Skip to main content

Full text of "Computational Modeling and Mathematics Applied to the Physical Sciences"

See other formats


NAb 
NAE 
DM 



CARNEQIE: 
PITTSBURGH, PA 15213-3690 



Computational Modeling 
and Mathematics 
Applied to 
the Physical Sciences 



Committee on the Applications of Mathematics 

Office of Mathematical Sciences 

Commission on Physical Sciences, Mathematics, and Resources 

National Research Council 



NATIONAL ACADEMY PRESS 
Washington, D.C. 1984 



NOTICE: The project that is the subject of this report was approved by 
the Governing Board of the National Research Council, whose members 
are drawn from the councils of the National Academy of Sciences, the 
National Academy of Engineering, and the Institute of Medicine. The 
members of the committee responsible for the report were chosen for their 
special competences and with regard for appropriate balance. 

This report has been reviewed by a group other than the authors ac- 
cording to procedures approved by a Report Review Committee consisting 
of members of the National Academy of Sciences, the National Academy 
of Engineering, and the Institute of Medicine. 



The National Research Council was established by the National 
Academy of Sciences in 1916 to associate the broad community of science 
and technology with the Academy's purposes of furthering knowledge and 
of advising the federal government. The Council operates in accordance 
with general policies determined by the Academy under the authority of 
its congressional charter of 1863, which establishes the Academy as a 
private, nonprofit, self-governing membership corporation. The Council 
has become the principal operating agency of both the National Academy 
of Sciences and the National Academy of Engineering in the conduct of 
their sendees to the government, the public, and the scientific and en- 
gineering communities. It is administered jointly by both Academies and 
the Institute of Medicine. The National Academy of Engineering and the 
Institute of Medicine were established in 1964 and 1970, respectively, un- 
der the charter of the National Academy of Sciences. 



This work was sponsored by the U.S. Department of Energy under 
contract number DE FG1081ER12001. 



Available from 

Office of Mathematical Sciences 
National Research Council 
2101 Constitution Avenue, NW 
Washington, DC 20418 



COMMITTEE ON APPLICATIONS OF MATHEMATICS 



WERNER C. RHEINBOLDT, University of Pittsburgh, Chairman 

JAMES G. GLIMM, New York University 

JAMES M. HYMAN, Los Alamos National Laboratory 

ROBERT J. KEE, JR., Sandia National Laboratories 

JAMES A. KRUMHANSL, Cornell University 

JOHN E. OSBORN, University of Maryland 

MARTIN H. SCHULTZ, Yale University 

IVAR STAKGOLD, University of Delaware 



JACOB K. GOLDHABER, Executive Secretary 



The Committee on Applications of Mathematics is a standing committee of 
the Commission on Physical Sciences, Mathematics, and Resources of the 
National Research Council. It is charged with taking all actions deemed 
appropriate to enhance the effectiveness of mathematics and mathematical 
statistics in their applications. 



m 



i-n*if%f 



COMMISSION ON PHYSICAL SCIENCES, MATHEMATICS, AND RESOURCI 



HERBERT FRIEDMAN, National Research Council, Chairman 

ELKAN R. BLOUT, Harvard Medical School 

WILLIAM BROWDER, Princeton University 

BERNARD F. BURKE, Massachusetts Institute of Technology 

HERMAN CHERNOFF, Massachusetts Institute of Technology 

MILDRED S. DRESSELHAUS, Massachusetts Institute of Technology 

WALTER R. ECKELMANN, Sohio Petroleum Company 

JOSEPH L. FISHER, Office of the Governor, Commonwealth of Virginia 

JAMES C. FLETCHER, University of Pittsburgh 

WILLIAM A. FOWLER, California Institute of Technology 

GERHART FRIEDLANDER, Brookhaven National Laboratory 

EDWARD A. FRIEMAN, Science Applications, Inc. 

EDWARD D. GOLDBERG, Scripps Institution of Oceanography 

CHARLES L. HOSLER, JR., Pennsylvania State University 

KONRAD B. KRAUSKOPF, Stanford University 

CHARLES J. MANKIN, Oklahoma Geological Survey 

WALTER H. MUNK, University of California, San Diego 

GEORGE E. PAKE, Xerox Research Center 

ROBERT E. SIEVERS, University of Colorado 

HOWARD E. SIMMONS, JR., E. I. du Pont de Nemours & Co., Inc. 

JOHN D. SPENGLER, Harvard School of Public Health 

HATTEN S. YODER, JR., Carnegie Institution of Washington 



RAPHAEL G. KASPER, Executive Director 



IV 



CONTENTS 



Overview 1 

1. Introduction 3 

2. Applications 8 

2.1 Hydrodynamic Systems 8 

2.2 Chemical Systems and Combustion 17 

2.3 Plasma Physics 24 

2.4 Particle Physics 26 

2.5 Condensed-Matter Physics 27 

2.6 Geophysical Applications 31 

2.7 Meteorology 33 

2.8 Astrophysical Applications 35 

2.9 Structural Mechanics 37 

2.10 Nondestructive Testing and Tomographic Reconstruction 38 

2.11 Mathematical Models in the Biological Sciences 41 

2.12 Electronic Components 45 

3. Computational and Mathematical Difficulties 47 

3.1 Degrees of Freedom 47 

3.2 Different Time and Length Scales 49 

3.3 Singularities in Coefficients, Data, or States 52 

3.4 Boundary Conditions 52 

3.5 Bifurcations and Chaos 54 

3.6 El-Posed Inverse Problems 59 

3.7 Effective Media 61 

3.8 Validation, Error Assessments, and Sensitivity Analysis 63 

4. Numerical Methods 67 

4.1 Discretization Methods 67 

4.2 Adaptive Grid Methods 72 

4.3 Methods for Solving Discretized Differential Equations 75 

4.4 Continuation and Homotopy Methods 78 

4.5 Optimization Methods 80 

4.6 Graph-Theoretical Methods 81 

4.7 Splitting Methods and Defect Corrections 84 

4.8 Monte Carlo Techniques 85 

4.9 Problem-Dependent Methods 89 

4.10 Computer Software and Hardware 93 
Appendix A: Persons Contributing to this Report 96 



OVERVIEW 



The 1980s are a time of profound challenge to the technological strength 
of the United States in the economic as well as the military spheres, and 
our country's performance in research and development in its engineering 
laboratories will be an important and perhaps determining aspect of our 
success in meeting this challenge. Advanced engineering development is 
now based mainly on scientific computation, which in turn relies on math- 
ematical modeling and laboratory experiments. Together they represent 
one of the areas in which the strength of nations is being tested today. 

Mathematics is essential in the development of theoretical and com- 
putational models for solving the highly complex problems of engineering 
and basic science, which encompass a range of scientific difficulties. On one 
side are questions of computer architecture and the science of algorithms. 
On the other side is the modeling of chemical and physical processes 
by means of mathematical equations. The issues are tied together by 
mathematical theory, which seeks a full understanding of the nonlinear 
phenomena contained in the equations and implements this understanding 
through computational methods. This span of scientific activities forms 
the subject that is known as applied mathematics. 

The principal conclusion of this committee is that computational 
modeling, which is a high-leverage element of our nation's scientific and 
technological effort, requires increased emphasis and support. The conclu- 
sion is documented by an examination of typical application areas, which 
reveals the pervasive difficulties that accompany computation of realis- 
tic problems and leads one to consider both what computers can do and 
what they cannot currently do but might eventually be capable of doing. 
As illustration, we examine several deep theoretical problems, including 
turbulence and combustion. At the frontiers of attack on these prob- 
lems we discover the limitations imposed by our current understanding of 
model formulation and computational capability. We examine modeling 
problems and algorithms adapted both to specific features of the desired 
solution and to the computer architecture. We also examine computer 
architecture and component design and manufacture as a mathematical 
modeling problem. 

The Committee recommends increased support for 

1. Research in computational modeling and applied mathematics, 

1 



2. Computing facilities dedicated to this area, 

3. Education and manpower development in computational and ap- 
plied mathematics. 

These recommended increases include financial support from govern- 
ment and industry as well as institutional support from universities. 



1. INTRODUCTION 



The extensive use of computers in advanced development work began dur- 
ing World War n, and today computing is a vital component of science, 
engineering, and modern technology. Most advanced technological devel- 
opment, from aircraft design to automobiles to petroleum to satellites now 
follows this pattern of reliance on the computer. Moreover, the needs of 
national defense have posed scientific and engineering design problems as 
difficult as any ever encountered. Numerical computation and applied 
mathematics have played an essential role in dealing with such problems. 
In fact, numerical fluid dynamics was born during the 1940s for the pur- 
pose of assisting in the design of nuclear weapons, combat aircraft, and 
conventional ordnance and is now applied widely by industry. 

Most problems of engineering or scientific interest are too complex 
to be modeled and computed exactly. Instead one considers a series 
of approximate models and computations, each of which illuminates a 
different aspect or idealized portion of the overall problem. When used by 
a skilled engineer or scientist, these mathematical models greatly enhance 
the judgment that goes into design decisions and reduce the amount of 
expensive laboratory and field testing required. These advantages account 
for the widespread use of these models. 

More specifically, mathematical models are used in engineering design 
problems in the following modes: 



1. To provide the first test of a new design idea. Beyond common 
sense and simple hand calculations, the computer model is usually the 
cheapest and fastest test of an idea. This test is applied before deciding 
whether to conduct a series of experiments or to build a prototype. 

2. To reduce the time and cost associated with laboratory and 
field tests. Usually engineering problems contain several critical design 
parameters that will have involved a certain amount of trial and error in 
the search for the optimum choice. The computer is used qualitatively 
(Will an increase hi parameter X improve or degrade performance - or 
performance parameter Y"?) and quantitatively (Which values of design 
parameters Xi,.. .,X n will optimize performance?). 

3. To assist in laboratory or field tests that determine model 
parameters and equations. Often the model parameters are measured 



only indirectly. Thus, a computer model of the laboratory apparatus 
may be needed to extract the desired information from the observed data. 
Usually these models are simpler than the complex engineering models, 
and their defining equations can be solved with greater precision and fewer 
approximations. 

4. To replace laboratory or field tests. Sometimes tests are impossible 
or impracticable. For example, measurement of chemical reaction rates 
at extreme conditions of pressure and temperature is very difficult, and 
accumulated experience through trial and error is not adequate for solving 
the problem of landing the Space Shuttle, for reasons of human safety as 
well as cost. 

5. To improve the education and judgment of engineers and scien- 
tists using the models. The mathematical models and computer solutions 
provide a vast increase in the quality and caliber of the data. Thus while 
the laboratory measurement may produce some overall quantity (e.g., total 
flow in and out), the computer model might yield detailed velocities and 
concentrations at each point of the flow field. Because the equations are 
nonlinear, it is difficult to foresee all the relevant phenomena, much less to 
understand then* relative importance. Thus the model becomes for users 
an experimental tool that allows them to understand a problem at a level 
of detail that cannot be achieved by other means. 



In summary, mathematical and computer models are used because 
they are faster, cheaper, and more effective. 

However, models have limitations. There are limitations in the 
validity of the equations used, in the adequacy of the solution algorithm, 
and in the size and speed of the computer. Also, the cost, accessibility, and 
reliability of computer software and, sometimes, the cost of the computa- 
tion itself can be limiting factors. These limitations in some sense define 
the frontiers of science, but more specifically they define the frontiers of 
applied mathematical science. 

Problems of realistic interest typically involve the study of diverse 
physical phenomena oA many scales of length in fully three-dimensional 
settings. Though essential, experimental science in these contexts is ex- 
pensive and difficult. The design of modern strategic weapons systems 
epitomizes these characteristics. For example, in the design of a TRIDENT 
submarine, the architecture of the vessel, the design of the missile, and 
the design of the nuclear warhead all need to be modeled and integrated, 
which requires many thousands of hours of computer time and stretches 
available computing power and modeling techniques to their limits. 

Generally, design and evaluation of new kinds of defense systems, 



such as remotely controlled or robotic vehicles, will require analysis of 
entirely novel systems, using parallel computer architectures among other 
things. The needs of industry for technological advances are similar and 
are dominated by such basic concerns as pollution, depletion of resources, 
energy conservation, and efficient use of manpower. It seems safe to 
say that defense and industrial needs will continue to lead numerical 
computation and applied mathematics into new and challenging regimes. 

Depending on the complexity of the problem and the magnitude of the 
effort expended, models range from excellent to merely suggestive in their 
quality and usefulness. In all cases, improvement of computer modeling is 
one of the most promising avenues to improved technological performance 
by OUT nation. 

As one of the aims of this report, the Committee wants to show and 
emphasize that in the computational approaches to most of today's press- 
ing and challenging scientific and technological problems the mathematical 
aspects cannot and should not be considered in isolation. There is a unity 
among the various steps of the overall modeling process from the formula- 
tion of the physical problem to the construction of appropriate mathemati- 
cal models, the design of suitable numerical methods, their computational 
implementation, and, last but not least, the validation and interpretation 
of the computed results. In particular, the Committee wants to illustrate 
that the steps are more often than not deeply interconnected and that 
the computational process may indeed be part of the model construction. 
At the same tune, there are problem areas, such as turbulence, where cur- 
rent theoretical research may promise a deeper insight into an important 
physical phenomenon. 

In line with these aims, the report uses a "matrix approach" that 
views the same problem from three different standpoints. In Chapter 2, the 
traditional approach is taken of discussing a number of typical problems 
leading to computational modeling from the viewpoint of the scientific 
or engineering discipline in which they arise. Then in Chapter 3 certain 
of these problems are touched on again, this tune from the viewpoint of 
the computational and mathematical difficulties that arise in connection 
with them. For example, these difficulties may involve large numbers of 
degrees of freedom, different scales of time and length, or singularities of 
various types. In Chapter 4 the viewpoint becomes that of the numerical 
algorithms involved in the computations, such as various discretization 
methods, continuation approaches, and splitting techniques. 

Of necessity, many topics have been left out. The list of applications, 
for example, is by no means complete, and, in fact, entire areas such 
as reactor safety and reactor physics are not mentioned at all. Neither 
did the Committee attempt to address all computational and mathemati- 



cal difficulties nor all variations of numerical algorithms. The choice of 
topics was motivated in part by energy-related considerations, the exper- 
tise and interests of the Committee and its advisers, and the report's broad 
purpose, whose achievement would be hindered by any attempt to be en- 
cyclopedic. For the same reason, the report is not intended to be a tech- 
nical summary, and this is also reflected in the fact that no attempt was 
made to reference the relevant literature. The report is mainly addressed 
to scientifically literate readers who know how to consult the literature 
when necesary. 

The Committee obtained advice and technical support from many 
colleagues across the country and from abroad. A list of names of all 
those who helped in this work is given in Appendix A, and the Committee 
is extremely grateful for all the often extensive documentations, special 
write-ups, and other comments that we received. The report is the result of 
a study begun by the Committee in 1981 on computational mathematical 
modeling and mathematics applied to the physical sciences with particular 
reference to the needs of the Department of Energy (DOE). The prepara- 
tion of the report was supported by the Applied Mathematical Sciences 
Research Program of the Office of Basic Energy Sciences of DOE, and the 
Committee also expresses its thanks and appreciation for this support. 

As stressed in the Overview, the Committee found that improvement 
of the mathematical and computer modeling of scientific problems is an 
important priority for our nation. The challenge is broad, and there are 
no simple remedies for current shortcomings. Accordingly, the Committee 
recommends the following: 



1. Increased research support for computational modeling and applied 
mathematics 

The technological challenges of the coming decades will impose new 
tests of our abilities in computational and applied mathematics, and meet- 
ing the tests will require increased research effort. As illustrated in this 
report, the challenges are typically multidisciplinary in nature with applied 
mathematics and modeling often in a central position. 

Hence, to support research in this area, multidisciplinary teams of 
adequate size to make progress on these complex problems should be 
encouraged; and organizational means should be devised to facilitate their 
establishment, continuity, and success. 



2. Increased support for computing facilities dedicated to computa- 
tional modeling and applied mathematics 

Ready access to modern computing systems is essential. Lack of 
equipment is the critical factor most strongly limiting academic research 
in computational mathematics. There is need both for conveniently usable 
local equipment and for access to large-scale computers. Local facilities 
are necessary for entire problems of modest size and for such tasks as 
code development, interactive debugging, test runs, and graphical analysis. 
Large-scale computing is essential because of the size of many typical 
problems as documented in this report. In this connection, the Committee 
strongly endorses the findings and recommendations in the recent Report 
of the Panel on Large-Scale Computing in Science and Engineering, Peter 
D. Lax, Chairman, National Science Foundation, December 26, 1982. 



S. Increased support for education and manpower development in 
computational and applied mathematics 

Today there are unmet manpower needs in computational and ap- 
plied mathematics, as discussed, for example, in Science and Engineering 
Education for the 80s and Beyond, a National Science Foundation report, 
October 1980. These needs are found in industry, government laboratories, 
and academic institutions. The critical challenges in this area call for a 
focus on quality. Specifically, the complex interdisciplinary nature of the 
problems poses special educational challenges for students and young re- 
searchers, and graduate and postdoctoral fellowship support for participa- 
tion in multidisciplinary teams of the type discussed above would also be 
helpful. 

A valuable aspect of such multidisciplinary education is the interac- 
tion it creates between applied mathematicians and other applied scientists 
in universities, government, and industry. 



2. APPLICATIONS 



2.1 HYDRODYNAMIC SYSTEMS 

Hydrodynamic processes touch nearly every aspect of our lives. In a report 
even many times larger than the present one, we could not possibly discuss 
all hydrodynamics applications. Therefore, we have chosen a few, with 
the hope that they will serve to illustrate the importance of hydrodynamic 
models and to point out some of the problems that occur in developing 
the models. 

A large fraction of the current research in computational modeling 
is done with hydrodynamics applications in mind. The great variety of 
responses that we received to our requests for material for this report 
attest to the diversity of these applications. Among them were applications 
having to do with aircraft and wing design, both at subsonic and supersonic 
speeds; global weather prediction and local phenomena such as tornadoes; 
water waves and ship hull design; piping networks, such as in nuclear 
reactor or power plant design; geologic phenomena, such as glacier flow 
or convection hi the Earth's mantle; biological flows, such as the flow of 
blood in the heart; and chemically reacting flow, such as combustion. 

The general system of equations governing hydrodynamics are called 
the Navier-Stokes equations. They are a statement of mass and momen- 
tum conservation, the momentum equation being a formulation of 
Newton's second law, F = ma. The Navier-Stokes equations were first 
developed in France by Navier in the early 1800s. They represented an 
improvement over the Euler equations that were first derived in 1755, in 
that the Navier-Stokes equations included viscous effects that were absent 
in the Euler equations. However, it was not until 1904, when Prandtl de- 
veloped the boundary-layer approximations, that predictions of practical 
viscous flows could be made. Practical solution of the full Navier-Stokes 
equations had to await the development of modern high-speed computers 
beginning in the 1940s. 

Why are equations that have been known for over 100 years so hard to 
solve? The answer lies largely in their inherent nonlinear characteristics. 
Immediately upon looking at the equations one sees that the convective 
transport terms (the acceleration in F = ma) involve velocity times its 

8 



gradient. This nonlinearity is always present, and it is responsible for 
the existence of complex phenomena such as shock waves and turbulence. 
In principle the Navier-Stokes equations alone provide a description of 
turbulence; however, one would have to resolve such small length scales in 
their solution that this approach is not of practical importance. Therefore, 
a great many approaches to approximating turbulence effects are being 
pursued. Typically, these models introduce further nonlinearities into the 
system. 

Other problems that arise in the solution of hydrodynamic problems 
are related to the disparate time and length scales that must be resolved. 
In particular, convective transport is characterized by the fluid velocity, 
whereas pressure waves travel at the sound speed-typically orders of mag- 
nitude faster than fluid speeds. At the same tune the effects of diffusive 
processes (e.g., shearing stresses) are felt instantaneously throughout the 
flow. In some cases the fluid can react chemically. In those cases, the reac- 
tion rates display a strong nonlinear dependence on the fluid temperature. 
This introduces yet more characteristic scales into the models. 

Depending on characteristic parameters such as the Reynolds number, 
the solutions to the Navier-Stokes equations either can be smooth and 
steady or they can exhibit regular oscillations. Or they can be completely 
chaotic. It is clear that depending on the regime the appropriate solution 
procedures could be quite different. 

Finally a word about boundary conditions. In some problems the flow 
is enclosed, and hence boundary conditions are applied at the boundaries 
of the enclosure (e.g., an automobile engine cylinder), which can often 
have a complex shape. In other cases, such as an aircraft wing, the flow 
is effectively unbounded, and the boundary condition should be applied at 
"infinity" (see Section 3.4). Some modeling problems arise in approximat- 
ing infinity by some finite boundary. In still other problems, such as 
the flow of blood in a heart, the boundary is both complex in shape and 
deforms depending on the forces exerted on it by the fluid. Modeling such 
a problem is clearly a challenge. 



2.1.1 Wings, Wind Tunnels, and Computers 

The economics of the energy shortage implies that planes will fly at speeds 
close to but less than the speed of sound. At such speeds there is a 
"supersonic bubble" over the wing, where the local velocity of the air 
relative to the whig is greater than the speed of sound. In this case the 
presence of shock waves is typical and undesirable. They are undesirable 



/\ / 


gfS 


I / 


t- J 


/ / / 


o g 


* * 


l S 


V 


s & 


** " 


1 p p 


J 


a * 


.* / 

/* / 


~ s? 


/ 


3 ^ 


1 


2 u 


( 


5^ 


\ 


n cd 




7 1 s 


1 1 ill 1 . 1 . 


OH fl> 

5 "^ 
m C^4 


09' 1- MM- Of- 1 Oh'- 00-'- Oh- OS' ; l 


CO t-1 


a 


13 " 
O -4^ 

;| g 




Q oQ 








q) O 




o. "^ 
^^ 


Xl \ 


J "CO 

' Id 

* 


+* " 


\ o3 O 


} 

t 


a a 


/ 


o o5 


I 


rt jj '*r n 


/ 


? a e c^ 1 

g > C 


I :' 


~. p) 2 


I i 


(=1 S cJ* 


I i 


^ : ^S^ 




^ 3 3 

J l!S 




U "rf 

4? *"* 




D M \ 


Oh 1 - 00'- Ok- U' 0-I 


g qs o 


dO 


O m 

a .a o 



/ 



c 



10 



because the drag can be computed as being proportional to the third power 
of the shock strength. The goal of efficient wing design is to produce 
wing shapes with no shocks or only weak shocks in this transonic region. 
A general mathematical theory shows that shockless wing foils exist for 
given transonic cruising speeds. However, the problem of finding such wing 
shapes is both overdetennined and extremely sensitive to small changes in 
the data, i.e., "ill-posed" (see Section 3.6). The solution to such ill-posed 
problems is still valid from an engineering point of view because operation 
at neighboring off-design conditions will produce only weak shocks and 
small drag, see Fig. 2.1. 

Using computer algorithms created by applied mathematicians, it is 
now possible to solve both the inverse problem (design) and the forward 
problem (of determining the flow field for a given wing shape and velocity) 
with sufficient accuracy that the use of costly wind-tunnel experiments can 
be greatly reduced. This accomplishment is a striking success of recent 
applied mathematics. 

The theoretical areas that have contributed to this study include the 
theory of nonlinear elliptic equations, complex function theory, and mixed 
problems, for which a prototype is the Tricomi equation, 



which is elliptic for y > and hyperbolic for y<Q. The elliptic region 
corresponds to the subsonic region, and the hyperbolic region to the su- 
personic "bubble." In a numerical method for the design problem, an 
analytic continuation makes the elliptic region hyperbolic. The resulting 
equation is solved by the method of characteristics, and then the analytic 
continuation back to real values of space is performed numerically. 



2.1.2 Chaos, Turbulence, and Droplets 

Turbulence produces a boundary layer along the trailing edge of an aircraft 
wing. The boundary layer degrades the wing performance and thus is an 
important part of the design problem. The flow behind the trailing edge of 
a wing contains a vortex sheet, and the roll-up of this vortex sheet produces 
turbulence that constitutes a safety hazard for small aircraft flying in the 
wake of a jumbo jet. The axis of the vortex roll-up is perpendicular to the 
wing, and so the roll-up is intrinsically three dimensional. In the simpler 
case of two-dimensional flow the vortex sheet is a line or curve in the 
plane, emanating from the trailing edge of the wing foil. In regions in 

11 



which the line is stretching, it is geometrically stable. In regions in which 
it is contracting, it is unstable. Instead of contracting, it forms a spiral 
vortex structure and hence is stable. 

The two-phase flow of water and steam in a cooling pipe, or of oil 
and gas in an oil-reservoir production well, is also a problem in which 
the geometrical instabilities of large-scale fluid motion are important. 
Here an internal movable boundary separates regions of different material 
properties. In some cases (a heavy fluid, e.g., water, falling into a light 
fluid, e.g., air), the material interface is unstable against formation of 
fingers. There is continuation of the nonlinear instabilities leading to 
pinchoff of droplets and a chaotic regime (mist) that can be analyzed on 
various length scales, as discussed below in the case of turbulence. 

During the combustion stroke of an automobile engine, the flame 
is quenched when it reaches the cold cylinder walls, and incompletely 
burned fuel present in the combustion chamber at this time contributes to 
pollution and to a loss of fuel efficiency. The rate of advance of the flame 
from the spark plug to the cylinder walls is governed by the laminar flame 
speed and the rate of turbulent mixing. Of these two effects, the second 
is more important. The turbulent mixing is produced by vortices that 
detach from the turbulent boundary layer at the wall during the intake 
and compression strokes (see Fig. 2.2). Thus, an accurate modeling of 
this problem requires an ability to treat a number of fluid singularities: 
flame fronts, vortices, turbulence, boundary layers, and boundary-layer 
separation. 

The examples above show that singularities in fluid flow may be 
geometrically unstable. When this instability occurs in a regime governed 
by the scale-invariant Euler equations, the phenomenon is repeated on 
all length scales and leads to chaotic solutions. Turbulence, vortex roll- 
up, convection fingering, and droplet formation are examples of this 
phenomenon, which we now discuss from a general point of view. 

The Euler equations of fluid dynamics allow intrinsic singularities, 
namely vortices, boundary and shear discontinuity layers, contact or 
material interface discontinuities, and shock waves (Fig. 2.3 shows a com- 
putation of the stretching of a vortex tube in a periodic inviscid flow). 
Depending on the problem, special discontinuities such as flame fronts or 
chemical reaction fronts (for fluid dynamics with chemistry) may occur. 
Within the singularity, the Euler equations fail to be a correct description 
of nature, and corrections (either parabolic effects or perhaps a more com- 
plicated Euler equation with more state variables) may be required. As an 
example, consider a shear layer (i.e., a jump discontinuity in the tangential 
velocity component). Taking the curl of the Navier-Stokes equations, we 
obtain for 

12 



= V X v = vorticity 



the equation 



+ (v V)u; (u 
at 



where v is the kinematic viscosity. To understand the significance of this 
equation, we specialize to two dimensions. Then w is a vector in the z 
direction, (w V)i>, and 



is the total, or Lagrangian time derivative, so that the Navier-Stokes 
equation says that vorticity moves by passive transport plus diffusion. 
The extra term, (a; V)v above, induces vortex production as a result of 
the stretching of vortex lines in three dimensions and is important for 
considerations of geometrical stability as discussed below. In summary, 
the diffusion term v&v of the Navier-Stokes equation is responsible for 
the vorticity leaving a boundary or internal shear layer and diffusing into 
the rest of the flow. Without viscosity there is no mechanism for vorticity 
to enter (an initially irrotational) flow region. The Prandtl boundary-layer 




FIGURE 2.2 Stretching of a flame by a vortical structure. Such stretching 
is important for the efficient operation of engines; it enhances burning by 
increasing the area of the flame. [From A. J. Chorin, Flame advection and 
propagation algorithms, J. Comput. Phys. 35, 1-11 (1980),] 



13 




al.p te aacnoni 1 to 



tip 30 "ittion 1 to 88 




atap E0 



"aUctlona 1 to IB 




step 



aic'Vlona 1 to 8 



FIGURE 2.3 Successive stages in the stretching of a vortex tube in an 
inviscid periodic flow. Vortex stretching is the mechanism by which energy 
in a turbulent flow is transferred to ever smaller scales where it is eventually 
dissipated, a, step 10, time = 0.65; b, step 20, tune = 0.88; c, step 30, 
time = 1.04; d, step 40, tune = 1.21. [Prom A. J. Chorin, The evolution 
of a turbulent vortex, Commun. Math. Phys. 83, 526-527 (1982).] 



14 



equations are a special version of the Navier-Stokes equations (scaled in the 
normal direction, so that diffusion occurs only normal to the boundary). 

Often the fluid singularities are geometrically unstable. They may 
bifurcate in a predictable fashion, developing "rolls" (Couette flow) or 
"cells" (Benard flow), or they may become irregular and highly convoluted 
with a tendency toward chaos, known as turbulence. There is no scientific 
reason to question the validity of the Navier-Stokes equation as a micro- 
scopic description of physics even into the turbulent regime. However, its 
usefulness as a description of large-scale fluid motions in the turbulent 
region can be questioned, and some other description of fluid flow, such as 
a random ensemble of interacting vortices, could be more effective. 

There are three energy or length scales in which quite distinct charac- 
teristic phenomena dominate. The smallest length scale is that on which 
energy dissipation dominates. On this length scale, the Navier-Stokes 
equations are the correct equations. The viscosity is large, causing velocity 
fluctuations to be rapidly smoothed and solutions to be (locally) "laminar." 
Since smoothness of solutions is a local property, we may conjecture that 
all solutions of the Navier-Stokes equations should be smooth for all time. 
This conjecture is the major unsolved problem of the energy dissipation 
range. It is known that solutions with smooth data will remain smooth 
for a short time, and solutions with smooth and small data will remain 
smooth for all time. Both statements exclude turbulent regimes. 

For weak solutions, Leray showed that the set of tunes t for which 
v(x, t) fails to be smooth is a set of measure zero. Recently considerable 
progress has been made in restricting the possible singularities of the 
Navier-Stokes equations. 

For longer length scales, viscous effects do not play a major role 
and the fluid flow can be described by Euler equations. However, this 
simplification gives rise to problems. The problems are not merely techni- 
cal but reflect the intrinsically complex phenomenology of fluid dynamics. 
The Euler equations are scale-invariant. Thus if v = v(x, t} is a solution, so 
is v = v(ax, at). The inertia! range is the set of length scales dominated by 
scale-invariant, universal physics. Whatever phenomena can occur (e.g., 
vortices) will be repeated on all length scales in the inertia! range. The 
inertia! range is limited at the smaller end by viscous dissipation. At the 
larger end, it is limited by the special boundary and initial conditions 
imposed on the flow, which result in special (nonuniversal) flow behavior. 

The inertial region is dominated by scale-invariant behavior. There 
is a flow of energy from large-scale motions to smaller ones (the "energy 
cascade"). This cascade seems plausible on physical grounds as a type of 
third law of thermodynamics but does not have a rigorous mathematical 
status. It can be explained as a consequence of the geometrical instability 

15 



of vortex lines and shear layers. As these go unstable, they generate 
(smaller) new vortices and vorticity. 

The energy cascade leads to a dimensional analysis of characteristic 
exponents and to the Kolmogoroff 5/3 power law 



for the energy distribution as a function of wave number fc. A discussion of 
the experimental data in connection with the Kolmogoroff theory has the 
vortices, which occur on all length scales in the inertia! range, as filling 
space. Actually, it may be better to assume the contrary: vortices of 
a given size fill only a small part of space. Then the smaller vortices, 
which are driven by the larger ones, will occur only within the region of 
these larger vortices, and in fact only within a small part of this region. 
This is the notion of intermittency. It leads to the idea of a singular set 
for solutions of the Euler equations, which is a Cantor set of fractional 
dimension less than 3. 

Intermittency leads to modifications in the Kolmogoroff exponent 
and to a renonnalization group type picture of turbulence. Numerical 
calculations to determine intermittency and energy cascade exponents 
have been performed. The calculations start by tracking vortex lines in 
a fluid flow and proceed through a sequence or renonnalization group 
length scale transformations to focus on the singular Cantor set within 
the solution. 

In most problems, the inertia! region contains lengths too small to be 
used directly in a fluid calculation. Its importance lies hi its role of fixing 
parameters such as an effective or eddy viscosity, which are then used to 
determine the large-scale motion of the fluid. The inertia! region is not 
particularly well understood from either the theoretical or numerical point 
of view. 

The large-scale fluid motions are produced directly by the initial and 
boundary conditions imposed on the flow. These motions are strongly 
problem dependent. Numerical calculations and experiment are important 
tools in their study as is the detailed analysis of simplified and idealized 
flow configurations. An important theoretical question is the evolution of 
initially unstable flow configurations. This question arises in connection 
with the onset of turbulence and in connection with the energy cascade, 
where large-scale vortices excite and drive small-scale ones. 

In some problems (supercritical turbulence), the instability in an ini- 
tially laminar flow is nonturbulent but arises from the bifurcation of a 
fixed point. Further bifurcations lead to higher-dimensional tori, and a 
general theory explains that generically the flow on the (sufficiently high- 
dimensional) torus has a strange attractor as its limit set and that this 

16 



strange attractor is chaotic in nature. This picture has been analyzed in 
the context of the Lorentz flow, which is the truncation of the Navier- 
Stokes equations to include only a small number of modes. The strange 
attractors found there have a Cantor-like structure. An example of super- 
critical turbulence is Couette flow. 

Subcritical turbulence occurs when the finite (noninfinitesimal) per- 
turbation is less stable than the infinitesimal one. Then turbulence occurs 
below the critical Reynolds numbers at which the linear theory shows 
instability and may go directly to turbulent behavior without a discrete 
sequence of nonturbulent bifurcations starting with laminar flow. 



2.2 CHEMICAL SYSTEMS AND COMBUSTION 

Prom the invention and manufacture of an enormous range of syn- 
thetic materials (e.g., plastics) to the refining and burning of fossil fuels, 
chemistry and chemical processes affect nearly every phase of our lives. 
Naturally it is important to understand and control, as fully as possible, 
many of these complex chemical processes. For example, we seek to find 
new and better materials, to reduce costs, and to generate energy more 
efficiently and with less pollution. Applied mathematics and computa- 
tional modeling continue to play a valuable role in meeting these goals. 

One of the oldest chemical processes harnessed by man is combustion. 
The successful modeling of combustion provides an extraordinarily rich 
source of challenges for the computational mathematician. Frequently, 
combustion models have to incorporate all the difficulties of complicated 
fluid mechanics coupled with complex chemical kinetics. The challenges 
include developing algorithms to ensure accuracy and to reduce computer 
tune and storage. The modeler also seeks appropriate simplifications that 
take advantage of any special attributes of the governing physics in order 
to gain more efficient computation. Since combustion contains a wide 
range of chemical processes, we use it here as an example to illustrate 
points of mathematical interest in general chemical systems. 

Even within the topic of combustion there is an enormous diversity 
of applications. The first topic that probably comes to mind is the model- 
ing of internal combustion engines, and this is an important application. 
Modeling is a part of ongoing research to design new types of engines (e.g., 
direct injected stratified charge), to improve fuel economy, to utilize alter- 
nate fuels (e.g., alcohols), and to reduce pollutant formation. Similarly, 
research for other combustors, such as gas turbines or power-plant boilers 

17 



benefit from computational models. Still there are many other important 
combustion problems aside from power generation. For example, the field 
of fire research is devoted to problems such as how fires spread in build- 
ings and the behavior of various fire-retardant materials. An important 
current topic in reactor safety is the characterization of hydrogen-air fires 
such as occurred in the Three Mile Island accident. Another public safety 
question deals with the problems of fire and explosion associated with 
a liquid-natural-gas tanker accident. Problems of burning coal and coal 
gasification are also topics of great current interest. 

Perhaps the simplest chemical process from a physical viewpoint is 
chemical equilibrium. At equilibrium all chemical reactions are assumed 
to have gone to completion, and the species concentrations are such that 
the mixture is in a minimum free-energy state. The mathematical com- 
putation of the chemical equilibrium state is posed as a constrained mini- 
mization problem. In combustion the equilibrium composition corresponds 
to the products of combustion long after the combustion is complete. 
Physically, the next step hi complication comes with the inclusion of finite- 
rate chemical reactions. Here the mathematical problem is one of solving 
systems of stiff ordinary differential equations (see Section 3.2 for a discus- 
sion of stiffness). Models of shock tubes or flow reactors, which are used 
frequently to probe fundamental questions in chemical reaction kinetics, 
fall into this class of problems. The physical situation is complicated fur- 
ther by the inclusion of fluid motion and heat and mass transport. In this 
case the mathematical problem is one of solving systems of parabolic or 
elliptic partial differential equations. 

Consider the internal combustion engine as an example. What are 
the things that we might hope to learn from modeling? Ultimately, we 
hope to influence geometrical considerations such as combustion chamber 
shape and component placement (e.g., valves, spark plugs, fuel injectors). 
We also hope to provide a fundamental understanding, on the molecular 
level, about how fuels are oxidized and how pollutants are formed. With 
such understanding we can suggest ways to alter the combustion process 
to advantage. In the past only power and size were important considera- 
tions, and engine optimization could proceed mostly experimentally. Now, 
however, there are too many parameters to optimize simultaneously, and 
computational modeling is increasingly important. 

Because of limitations hi available computational resources, two tacks 
are taken in engine modeling. One is to consider mostly hydrodynamic 
effects. Here the modeling of boundary shapes and component placement 
is of primary importance (e.g., How should the piston face be shaped, 
and where should the spark plug be placed?). Complex domains and two- 
and three-dimensional effects are important. The models must incorporate 



18 




FIGURE 2.4 Velocity vectors computed in a direct-injection, stratified- 
charge engine at a position near top dead center. The combustion takes 
place in a swirling environment in a cup-like region machined into the 
piston. Several vortices are seen to develop in the cup. [Prom T. D. Butler, 
L. D. Cloutman, J. K. Dukowicz, and J. D. Ramshaw, Multidimensional 
numerical simulation of reactive flow in internal combustion engines, Prog. 
Energy Combust. 3d. 7, 293 (1981).] 

turbulent hydrodynamic effects and sometimes phase-change processes, 
such as fuel spray injections. The chemistry is usually simplified in these 
models because it is not feasible to consider both complex chemical kinetics 
and hydrodynamics on current computers. Figure 2.4 shows the velocity 
vectors that result from a two-dimensional simulation of a direct injected 
stratified charge engine. This is a new engine concept that is being studied 
in a U.S. Department of Energy-sponsored cooperative program including 
General Motors Research Laboratories, Princeton University, and three 
National Laboratories. 



19 



In addition to the hydrodynamic aspects of engine combustion, there 
are important unanswered questions about the chemistry. Therefore, the 
second tack is to consider simplified hydrodynamic situations, such as 
laminar flames, and treat the chemical kinetics in great detail. These 
models address issues such as ignition phenomena and pollutant formation. 
Figure 2.5 shows some species profiles computed in an atmospheric pres- 
sure acetylene-oxygen premised flame (acetylene is an important reactant 
in soot formation). This model used 30 chemical species and 103 reactions. 
The results were computed using an adaptive mesh placement strategy, 
and they resolve detailed structure within the flame. Note that the flame 
is very "thin"- its thickness is on the order of one millimeter, while com- 
bustion chamber dimensions are on the order of tens of centimeters. 

An interesting aside is to note that these two approaches to combus- 
tion modeling match corresponding approaches to experimental investiga- 
tion. That is to say, it is usually not possible to measure or compute minor 
species concentrations in complex turbulent flows, whereas it is possible 
to do so in laboratory flames. 




0.00 0.02 0.04 0.06 0.08 0.10 0.12 0.14 

X(CM) 



0.16 



FIGURE 2.5 Species mole fraction profiles showing the internal structure 
of an atmospheric-pressure stoichiometric acetylene-air flame. [From J. A. 
Miller and R. J. Kee, Sandia National Laboratories.] 



20 



The inherently disparate time scales of chemical reactions, in other 
words their stiffness, contribute to the numerical difficulty of solving com- 
bustion models (see Section 3.2). The inherently disparate tune and length 
scales for fluid transport, heat transfer, and chemical reaction are respon- 
sible for the presence of steep fronts in the solutions. Also, there are many 
degrees of freedom in the system of equations. The number of governing 
partial differential equations is large because a transport equation must be 
included for each species involved in the chemical reaction set. A system 
of 30 to 50 species, involving 100 or more chemical reactions is typical even 
for fuels as simple as methane. Also, for practical combustors, the model 
must ultimately include complex three-dimensional geometries. Because 
complete models of real combustors are too large for present computers, 
an important challenge is to simplify the models (including the physical 
assumptions) to a tractable level. In addition to these problems, there are 
potential difficulties related to scaling. The temperatures are on the order 
of 10 3 K, while some species can have important effects even when their 
mass fractions are as low as 10~ 10 . Moreover, before the computation, 
the peak mass fractions of the various species are usually known only to 
within several orders of magnitude. 

Many of the challenges of combustion modeling have been met by the 
numerical-analysis community; however, many more await resolution. For 
example, for systems of ordinary differential equations we understand how 
to treat the stiffness that results from the complex chemical kinetics by 
using stable implicit methods. However, for systems of partial differential 
equations the application of these methods leaves many open questions 
about how to treat the linear algebra and how to compute the error 
estimates. Operator splitting methods are important in rendering the 
linear algebra tractable for large problems. Stillness also occurs in low- 
Mach-number flows due to very high velocity, but low-amplitude, pressure 
waves. Usually, we do not care about the details of these waves, but 
they can unreasonably limit the size of the time step in explicit methods. 
Subcycling methods, rather than implicit methods, are sometimes used 
to alleviate this problem. Subcycling is a form of operator splitting (see 
Section 4.7) in which the invicid hydrodynamic equations are solved with 
small time steps, while the viscous parts of the equations and the energy 
and mass transport equations are solved using a much larger time step. 

Another particularly important topic in combustion models concerns 
the adequate resolution of localized behavior such as flame fronts. One 
line of research considers adaptive meshing strategies in which a spatial 
mesh network is adjusted dynamically so as to capture the local behavior 
accurately. Other work considers front-tracking methods, where the flame 



21 



is treated as a local discontinuity and conservation equations connect both 
sides of the front and predict its movement. 

Combustion models must account also for fluid turbulence, an area 
where even the underlying physics is not well understood. Here, computa- 
tional models such as the random vortex method are proving valuable 
in simulating turbulence effects, especially in the investigation of large- 
scale turbulent eddies, the so-called coherent structures. We classify these 
methods as "problem-dependent methods" because the physics and the 
numerical model depend heavily on each other (see Section 4.9). Recently, 
the random vortex method has been combined with a flame propagation 
algorithm so that combustion events can be modeled. Figure 2.6 is a se- 
quence of computer plots that shows the vortex velocity fields and flame 
fronts as computed from a model of turbulent combustion behind a step. 



2.2.1 Asymptotic Analysis 

We mention here that applied analysis (in concert with computation) is 
valuable for many problems hi combustion. Asymptotic methods, for 
example, can actually take advantage of phenomena such as steep fronts. 
They can thus be used to provide insights and to suggest approximations 
that help to reduce the complexity of the numerical models and thus render 
them more tractable. An important aspect of the asymptotic analysis is 
the possibility of considering the dynamical stability of flames. Such work 
may lead to fundamental understanding of such phenomena as the onset 
of turbulence. 

Asymptotic methods exploit the fact that the overall activation energy 
of the chemical reaction is typically large, a well-known consequence of 
which is that flame fronts are very thin. That is, chemical reaction is 
only important in a thin region where the temperature first approaches 
its burned value. On the unburned side of this region, chemical reaction 
is negligible because the temperature is too low, and on the burned side, 
it is negligible because the reaction has essentially gone to completion, 
depleting the unburaed fuel. Mathematically speaking, this thin chemi- 
,cally reactive region may be thought of as an internal boundary layer, 
separating the unburned and burned gases. In the limit, of asymptotically 
large activation energy, the boundary layer is infinitely thin, and we may 
use asymptotic matching principles to connect, or match, the solutions in- 
side and outside the boundary layer. The result is a flame sheet model in 
which the solutions on either side of the sheet are connected by nonlinear 
jump conditions that depend on local conditions at the front. Though the 

22 



resulting problem is still nonlinear, it represents a significant simplification 
over the original problem with Arrhenius kinetics and is in fact equivalent 
to the original problem in the asymptotic limit of infinitely large activation 
energy. 

The first studies using large activation energy asymptotics began to 
appear about a dozen years ago and were noteworthy for providing for the 
first time analytical flame speed formulas for a steadily propagating planar 
flame. The most significant results to come out of this approach to date are 
the predictions of instability and bifurcation phenomena, which describe 
cellular and pulsating modes of flame propagation. These more complex 
solutions branch, or bifurcate, from the basic solution (in this case, a 




^7^^*"*""*r^"^^-o 









T^f* - 1 





FIGURE 2.6 A sequential series of computer plots displaying velocity 
fields and flame fronts in turbulent combustion behind a step at inlet 
Reynolds number of 10000. The combustion is stabilized by a recirculation 
region behind the step. The unburned gas is a mixture of propane and air 
at an equivalence ratio of 1/2. [From A. F. Ghoniem, A. J. Chorin, and 
A. K. Oppenheim, Numerical modelling of turbulent flow in a combustion 
tunnel, P/wl Trans, R. Soc. London. Ser. A 304, 303-325 (1982).] 

23 



steady, planar flame) as one or more parameters exceed critical values for 
instability of the basic solution. Their significance lies partly hi the fact 
that they represent intermediate modes of propagation of the flame in its 
evolution from a laminar to turbulent form of propagation. Experimental 
observations of these transitional modes in their pure form have recently 
been reported, and the theory thus identifies critical parameter thresholds 
that separate one form of flame propagation from another. A number 
of bifurcation analyses using nonlinear perturbation methods have been 
successful in characterizing these nonsteady and/or nonplanar flames for 
parameter values near the instability threshold. However, numerical ap- 
proaches may be used to great advantage in describing these transitional 
solutions when parameters far exceed their critical values for instability of 
the basic solution. 



2.3 PLASMA PHYSICS 

Controlled nuclear fusion provides a possible long-range energy source. 
As currently conceived, a fusion device must contain a deuterium and 
tritium plasma for a sufficiently long tune for net energy production. The 
confinement may be effected either by magnetic fields or by simple iner- 
tia effects. Inertia! confinement of a pellet, coupled with laser or particle 
beam heating, is a distinct possibility, although the bulk of the world 
fusion program centers on magnetic confinement of a plasma. The un- 
derlying problems in magnetic confinement are the determination of the 
equilibrium, stability, transport, and heating properties of plasmas under 
realistic conditions. Numerical modeling and computation represent major 
tools in this study. In recognition of their importance, the Magnetic Fusion 
Energy Computing Center has been established at Lawrence Livermore 
National Laboratory in Livermore, California. This Center is the third 
largest scientific computing center in the United States, and its sole func- 
tion is to provide computing capability to the scientists and engineers in 
the U.S. fusion community. Fusion plasma physics spans a diverse collec- 
tion of fields, with significant efforts occurring in the fields of Hamiltonian 
particle dynamics, statistical mechanics, kinetic theories, and dissipative 
and nondissipative single and multifluid models. The complexity of the 
problems involved makes computation an integral part of the research and 
development program for fusion. 

Controlled-fusion confinement experiments have indicated that, even 
in grossly stable configurations, fluctuations may play an important role in 
detennining energy and particle transport. Thus, an understanding of the 

24 



nonlinear behavior of various plasma models is fundamental to describing 
such behavior. For both ideal magnetohydrodynamics, a fluid model, and 
the Vlasov-Maxwell system, a kinetic model, linearized equations have 
been studied extensively. These analyses are appropriate for describing 
small-amplitude deviations from a quiescent equilibrium but omit the 
effects of mode-mode coupling and the onset, properties, and evolution 
of turbulence. The development of a self-consistent model of plasma 
equilibrium with fluctuations, stochastic particle and magnetic-field-line 
behavior, and resulting transport will continue to be an important research 
area. 

Plasma fusion applications present many problems in which the equa- 
tions of Hamiltonian dynamics appear. These systems describe single- 
particle phase-space trajectories in the Vlasov-Maxwell theory of colli- 
sionless plasmas. In addition, the trajectory of a magnetic field line in 
a toroidal system is described by a nonlinear Hamiltonian. Thus, the 
question of the existence and construction of adiabatic invariants, ex- 
plicitly time-dependent or not, is of fundamental importance for these sys- 
tems. Significant progress in this area might be made by combining ideas 
from modern topological dynamics, numerical simulation, and perturba- 
tion analysis. For example, if magnetic field lines are ergodic throughout 
a volume rather than lying on closed invariant surfaces, as is given by 
the Kolmorogoff-Arnold-Moser theory, then, owing to electron streaming, 
the thermal conductivity within this volume will be very fast. Thus, such 
questions as when ergodicity arises and what the properties of Hamiltonian 
dynamics are under ergodic circumstances impact strongly on these ap- 
plications. 

Since computational modeling of the full three-dimensional plasma 
equations is out of the question, current work utilizes a range of different 
compromises, simplifying to different degrees either the physics or the 
geometry to obtain a number of computationally tractable problems, each 
of which illuminates in a distinct fashion, a different aspect of the full 
plasma problem. In spite of these difficulties, the computational approach 
is a major route to progress on the problem of controlled fusion. The 
reason is simple. Experiments are expensive and must be supplemented to 
the maximum extent possible by theory. The theory is highly complex and 
nonlinear and is obtained by a combination of numerical experiments and 
physical intuition. Clearly improved computational methods are one of the 
methods through which progress is achieved in this area. Development of 
more powerful computers will also be required, as it is hard to believe that 
smart algorithms will by themselves suffice. 



25 



2.4 PARTICLE PHYSICS 

The problem is to find the equations that describe the elementary particles 
of subnuclear physics. This problem is among the most difficult to have 
been considered seriously by science hi this century. In recent years there 
have been significant gains in OUT understanding of the mathematical 
structure of the equations of quantum field theory. An analysis of the 
mathematical existence question for the quantum 4 equation 



shows that in space-time dimension d < 3, the theory exists, whereas in 
d > 5, there is no such theory (as would be given by standard methods) 
except for the trivial case X = 0. The method of proof suggests nonex- 
istence for the physical case d = 4 as well and offers confirmation from 
the side of exact mathematical analysis of ideas advanced by theoretical 
physics. This mathematical confirmation is much more compelling than 
any confirmation yet offered from experimental physics. 

The < 4 model is "near" to the Sine-Gordon equation, which is com- 
pletely integrable. Among recent results are the discoveries of a large 
class of stable, localized (in space), periodic (in time) solutions (analogs 
of the Sine-Gordon breather) and of the existence of breather formation 
resonances in soliton/antisoliton scattering. There are many related chal- 
lenging problems in nonlinear mathematics and dynamical systems theory. 

The problem with the 4 equation in d = 4 dimensions is one of several 
reasons for considering in its place quantum gauge fields. A standard 
approach to the quantization problem reduces it to the existence question 
for a singular non-Gaussian functional integral over an mfinite-dimensional 
space. In the case of gauge fields the gauge potentials form an infinite- 
dimensional afGne space A on which the group Q of gauge equivalence 
acts. The functional integral is only defined over the mfinite-dimensional 
manifold 



This manifold is not flat. The integration over M has been shown by exaxit 
mathematical analysis to require introduction of coordinate patches. Only 
locally, within a single coordinate patch can M AfS be regarded as an 
open subset of Euclidean (Hilbert) space. 

One of the methods proposed for understanding integration over M is 
to understand the critical points A 6 M. This question leads to a study of 
classical solutions of the Yang-Mills equation and a reduced form of this 
equation, called the self-dual Yang-Mills equation. Here we are looking at 

26 



a specific nonlinear elliptic equation in d = 4 dimensions. A remarkable 
analysis has led to a complete classification of its solutions, using methods 
of geometry, topology, and algebra (fiber bundles, the index theorem, and 
algebraic varieties). 

The development of numerical methods for the study of quantum 
fields is in its infancy, mainly because the problem is to a large extent out 
of range of present-day computing machines. Among the methods used 
on this problem, we mention Monte Carlo integration over a space-time 
discretized version of M = AfQ. The discretized problem is called a lattice 
gauge field, and even after discretization, the dimension of M is too high 
to allow evaluation of integrals by direct quadrature. A second approach is 
to generate series coefficients from cluster expansions of statistical physics 
and to reconstruct the desired function space integral from a Fade* analysis 
of the coefficients. Other attempts have been based on finite elements and 
on the renormalization group. It may be that a collaboration of numerical 
analysts with mathematical or theoretical physicists on this problem would 
be beneficial. 



2.5 CONDENSED-MATTER PHYSICS 



2.5.1 Statistical Physics 

In this subject, the equation of state, transport coefficients (pressure, 
viscosity, and thermal conductivity, for example), and other macroscopic 
properties of matter are related to and derived from intermolecular forces. 
The area has diverse and important applications, ranging from metallurgy 
to polymer chemistry to semiconductors and is an active area of research 
from the points of view of theory, numerical methods, and experiment. 
Here we limit ourselves to the two mathematical aspects of this subject: 
mathematical theory and numerical methods. 

The equations of statistical physics involve a large or infinite number 
of degrees of freedom, and so the mathematical theory of use here is 
analysis over infinite-dimensional spaces. Almost the same mathematical 
structure arises as in the study of quantum fields, the relationship between 
the two being that a quantum field is a continuum limit of a statistical 
physics (crystal or lattice) model. 

In a small class of models, including the two-dimensional Ising model, 
exact solutions are known. A larger but still restricted class of models has 
been analyzed mathematically with respect to qualitative behavior. One 

27 




3.43 



FIGURE 2.7 Monte Carlo simulation of a kinetic Ising model compared 
with x-ray and neutron deft-action measurements on an alloy of 60 percent 
gold and 40 percent platinum that has been heated and then quenched to 
60 percent of its critical temperature. The abcissa is a reduced momentum 
transfer, and the ordinate is a reduced scattering intensity. (From M. H. 
Kalos, New York University.) 



issue recently addressed hi this work was the stability of surfaces, as is 
relevant to the problem of crystal growth. Another is the effect of aperiodic 
or random crystal structure on bulk material properties. A striking recent 
development was the mathematical demonstration of the dipole binding 
(Kosterlitz-Thouless) phase transition. 

The numerical methods for statistical physics are basically of three 
types. In molecular-dynamics calculations, one takes a large number of 
particles and follows their motion by integration of the ordinary differential 
equation 



dt 2 l ' " '' 

defined by the mtermolecular forces F*. This method is "exact" in its 
treatment of mtermolecular forces, to the extent that they are known 
and that quantum effects can be neglected, but it is approximate in its 
treatment of statistics, since there are computer-dictated limits on the 
number N of particles that can be included. 

28 



Next we mention the method of series expansions, applied to the 
calculation of the equilibrium distribution dw and the partition Z, which 
typically has the form 



and 



z = 



where V is the intermolecular potential energy. Then thermodynamic 
functions such as pressure emerge as derivatives or logarithmic derivatives 
of Z with respect to the parameters (e.g., in V). If the potential energy 



is a sum of pair potentials U then the identity 



-4.50 
-5.00 
-5.50 
-600 
-6.50 
-7.00 
-7.50 



-8.00 




0.280 0.320 0.360 0.4OO 0.440 0.480 0.520 



FIGURE 2.8 Energy of liquid helium as a function of density. The broken 
line and points with errors show results of the Monte Carlo quadrature of 
the many-body nonrelativistic Schrodinger equation. The solid line is a fit 
to the experimental data. (From M. H. Kalos, New York University.) 



29 



substituted in the definition of du above yields an expansion that is rapidly 
convergent, where 

\e- u -l\l 

i.e., where L/~0. This is the region of noninteraction, and so series ex- 
pansions are especially useful to give weak corrections, for example, real 
gas corrections to an ideal gas. The expansions appear to converge up 
to critical points and with considerable work have been used to extract 
information on the equation of state in that region. 

Monte Carlo methods are used in the quadrature of very large- 
dimensional spaces, such as the determination of du above, as well as in 
the direct simulation of stochastic systems. These methods have become 
an experimental tool of mathematical physics. In studying the qualita- 
tive and quantitative behavior of a highly idealized model such as the 
Ising model in equilibrium or very far from equilibrium, or of lattice gauge 
theories, continuum and lattice models of polymers, and atomic models of 
quantum liquids and solids, one can carry out numerical studies in which 
the high-dimensional (i.e., many-body) character of the problem is not 
distorted. 

In application to the calculation of the equilibrium distribution du), 
the essence of the Monte Carlo method is as follows: Starting from an ar- 
bitrary point in a given ensemble, one modules a single particle position Xi 
"at random" but usually so as to lower the potential energy V, occasionally 
so as to raise V. After enough such elementary steps, convergence to the 
distribution duj is obtained. As described here, the method is extremely 
simple, and complications arise from the necessity to obtain convergence 
in a reasonable tune for realistic problems. 

Perhaps the most significant success has been in the microscopic 
theory of classical fluids, where Monte Carlo modeling has provided the 
"experimental" basis for the accurate expansion in \e~ u 1| mentioned 
above. Another conspicuous success is the extraction of critical exponents 
by joining ideas of the renormalization group to Monte Carlo simulation. 
Figure 2.7 shows Monte Carlo simulation of a kinetic Ising model compared 
with scattering measurements on a real alloy. 

Figure 2.8 shows how Monte Carlo methods applied to the many- 
body nonrelativistic Schrodinger equation give a quantitative account of 
the energy of (real) liquid helium as a function of density. 



30 



2.6 GEOPHYSICAL APPLICATIONS 

Out of a wide range of geophysical applications we limit ourselves here to 
three modeling problems in connection with the discovery and production 
of petroleum. Mathematically the discovery process is an inverse problem 
(see Section 3.6) that of constructing geologic maps using seismic signals, 
which in turn are generated by vibrations or explosions at the surface of 
the Earth. As the signals penetrate the Earth, they cross layers of differing 
density. This causes reflected signals to be returned to the surface, where 
they are then recorded and analyzed. The raw data are very noisy, owing 
to irrelevant near-surface density fluctuations. The noise is removed by 
averaging signals from neighboring receptors or from neighboring source 
locations. Then multiple reflections must be subtracted, and a compen- 
sation (called normal moveout) must be introduced for effects of nonver- 
tical signal propagation. Reflection from nonhorizontal layers generates 
complicating shear waves as well as pressure waves. The subtraction of 
multiple reflection signals can be based on Fourier analysis in a half-space 
and a Wiener-Hopf factorization. 

The correction for nonvertical signals can be based on a reduced 
Helmholtz equation, (-A + fc 2 + v)u - 0, but in order to focus on sig- 
nals moving in only one direction (either up or down), one takes first a 
square root and then a power series expansion of the square root. This 
process is known as the parabolic approximation and leads to the familiar 
Schrodinger equation. Alternatively, the analysis can be based on ray 
tracing and the solution of ordinary differential equations. The numerical 
problem is to implement these steps efficiently, in view of the large amount 
of data to be analyzed. 

The production problem is to describe the flow of oil, water, chemicals, 
and/or heat in a porous sandstone layer. The equations to be solved 
are a coupled system of nonlinear parabolic equations. Generally the 
equations describe mass conservation of individual species, and typically 
some are nearly elliptic whereas others are nearly hyperbolic. In the 
hyperbolic equations, coherent shock and rarefaction waves describing oil 
banks, absorption fronts, and flame fronts may form (see Sections 3.2 and 
3.3). Depending on a dimensionless number known as the mobility ratio, 
heterogeneities, and geometrical effects (convergence versus divergence), 
the fronts may become unstable with respect to the formation of fingers 
(see Section 2.1). 

Critical issues in this problem are the efficient solution of large sparse 
linear systems arising from space and time discretizations and the con- 
trol of numerical instabilities in the solution methods. The physical in- 

31 




FIGURE 2.9 Successive time steps in the movement of an oil-water in- 
terface toward a producing well. The view is a vertical cross section, 
with the well located on the left boundary of the computational region. 
The instability is caused by three factors: by a heterogeneous conducting 
channel near the bottom of the reservoir, by the converging cylindrical 
flow pattern near the producing well, and by the better flow properties of 
the less viscous displacing fluid (water). [From J. Glimin, B. Lindquist, 

0. McBryan, and S. Yaniv, Statistical fluid dynamics II: the influence of 
geometry on surface instabilities, in Frontiers in Applied Mathematics, Vol. 

1, R. Ewing, ed., SIAM, Philadelphia, Pa. (to be published).] 

stabilities mentioned above are controlled by the use of heat (to make 
the oil flow more easily) or the use of a heavier pushing fluid (polymer- 
thickened water in place of water or water and CO 2 in place of CC>2). 

The third modeling problem we discuss is the process of lifting the 
oil to the surface of the Earth. Because of the sudden drop in pressure, 
considerable gas will typically come out of solution, and the resulting two- 
phase mixture is in the droplet or mist regime mentioned in Section 2.1. 
In contrast to the case of reservoir flow, even the basic equations for two- 
phase flow in a pipe are not well understood. 



32 



2.7 METEOROLOGY 

Accurate forecasting of the movement of large-scale weather patterns is 
clearly an important problem. Here we mean the tracking of highs and 
lows on the order of 1000 kilometers in diameter. In addition, we are 
also concerned with the prediction of cloud patterns, precipitations, tem- 
perature, and other elements of the weather in as much detail as possible. 
Global predictions of the atmospheric flow are now made routinely at as 
many as 12 levels in the troposphere and stratosphere. 

By studying dynamic models of the atmosphere, estimates have been 
made of the sensitivity of the atmosphere to small perturbations, often 
quantified in terms of the error doubling tune. If two initial states of the 
atmosphere differ by random variations of 1, then it is found that the 
resultant states will differ by 2 in about 3 days and by about 4 in about 
6 days. Hence, meteorologists have said that the weather is unpredictable 
for periods of more than 2 to 3 weeks. But, the forecasters are still far 
from being able to predict the weather accurately for even 1 week. The 
best results are currently obtained by the European Centre for Medium 
Range Weather Forecasting, whose 4-5 day forecasts are currently superior 
to those of the U.S. National Weather Service. 

The accuracy of short-range predictions is limited by four somewhat 
distinct factors: 



1. Initializations: (a) The accuracy and completeness of observational 
data and (b) the compatibility of these data with the mathematical model, 
which is, of course, a simplified representation of the atmosphere. 

2. The limited number of grid points or spectral components used 
in the truncated, i.e., numerical, model as a result of limited computer 
power. 

3. Missing or severely truncated physical processes such as cloud 
dynamics, proper representation of turbulence, and radiation fields, lead- 
ing to inadequacy of the mathematical model describing the atmosphere. 

4. The inherent finite limits of predictability of certain types of non- 
linear dynamical systems such as the atmosphere. (Compare the previous 
paragraph.) 



The study of problems related to the reduction of the influence of 
these factors is at the heart of most research that is directly related to the 
improvement of weather prediction. 

The sparsity and inaccuracy of observational data [l(a)] are being 

33 



partially overcome by the use of weather-observing satellites. Concerning 
l(b), considerable research is currently directed toward the analysis and 
filtering of the observed data so as to make them compatible with numeri- 
cal models. Items 2 and 3 are closely related since, in a general way, the 
lack of a complete representation of all physical processes is related to the 
incomplete numerical resolution of the mathematical model. Such physi- 
cal processes include cloud dynamics (and the associated cloud physics), 
boundary-layer dynamics (including detailed features of the terrain and 
local heat, moisture, and dust sources), and detailed radioactive processes, 
In item 4 a new concept, "finite limits to predictability" has emerged 
from the work of E. N. Lorenz, namely the idea that the atmosphere and 
perhaps certain other dynamical systems may have finite limits of predic- 
tability regardless of the accuracy and detail of their initial conditions 
and the accuracy with which they are computed. Lorenz concluded from 
a simplified model that hydrodynamical systems with an energy spectrum 
(in the customary sense) having a power-law exponent greater than 3 
(e.g., -5/3) should be unpredictable in detail after a finite time. He sug- 
gests a time of about 15 days for the atmosphere, whose large-scale energy 
spectrum seems to have an exponent of about 2.8, with smaller scales 
(i.e., less than say, 20 km) having the 5/3 power law associated with 
homogeneous turbulence. This question of predictability is clearly related 
to that of the stability of hydrodynamical systems and to modern theories 
of bifurcation and chaos (see Section 3.5). 

A large component of atmospheric research is concerned with topics 
that have little direct application to forecasting. Their motivation may 
lie in basic scientific inquiry or in applications to other disciplines, e.g., 
radio-wave transmission for satellite communications. They may involve 
the study of fundamental hydrodynamic phenomena or other physical 
processes several steps removed from specific application. In all of these 
studies we find extensive use of analytical and numerical modeling of 
physical systems and processes. In addition, a wide variety of statistical 
approaches are used hi all areas of atmospheric research. 

Let us mention a few more areas of study. In addition to the predic- 
tion of large-scale weather patterns, computers are being used in limited 
regions to study the development and movement of smaller-scale distur- 
bances, such as hurricanes, thunderstorms, and tornadoes. Here adaptive 
grid methods are used (see Section 4.2). It is hoped, for example, that a 
better understanding of the details of these destructive phenomena will 
lead to the possibility of altering their course and development. 

On the longer time scale, where seasonal forecasts of average rainfall 
and temperature are made over large regions of the globe, there are at 
present several potentially useful, but unproven ideas. Here, however 



34 



some new mathematical and physical insight is needed in order to develop 
a satisfactory "average" system of equations. This is an important open 
problem. 

On the still longer time scale of decades and beyond, determining 
the effect of our mechanized civilization on the environment is a basic 
problem. Significant progress has already been made, through the use 
of simple climate models in conjunction with paleontology, astronomy, 
geology, and volcanology, in efforts to understand the factors that influence 
irregular alternations of ice ages and interglacial periods. These statistical- 
dynamical climate models will be essential for predicting the long-term 
effects of the observed increase in carbon dioxide and similar problems 
that will almost certainly arise as our industrial civilization expands. 



2.8 ASTROPHYSICAL APPLICATIONS 

Astrophysical studies pose a wide range of problems for mathemati- 
cal and computational analysis. The underlying physical theories in- 
clude hydrodynamics, magnetohydrodynamics and plasma theory, radia- 
tive transport theory, atomic and nuclear physics, and general relativity 
theory, and hence various comments made in other applications sections 
of this report apply to this area. 

In recent years considerable progress has been made in theoretical and 
numerical studies of a number of astrophysics! topics, as, for instance, the 
study of stellar interiors, the formation of stars and galaxies, the spiral 
structure of galaxies, the physics of supernovae and the evolution of su- 
pemovae remnants, the formation of the solar system, and the behavior of 
binary star systems, to mention just a few. The computational approaches 
may range from relatively simple simulations to the numerical solution of 
complex systems of partial differential equations. 

As an example of the first type, simulations have been applied to 
the study of star formation in a galaxy. It may be assumed that when a 
massive star becomes a supernova the shock wave emanating from it can 
compress the surrounding interstellar gas creating new stars. If at least 
one new star is also a massive star the phenomenon can repeat, leading 
to a chain reaction in the creation of stars. This is equivalent to a direct 
percolation problem, and, as is typical for such problems, phase transitions 
are involved. This is a nonlinear problem with a complicated structure in 
space and time. Accordingly, analytic techniques are difficult, whereas 



35 



computational simulations are relatively straightforward and lead to the 
development of realistic model galaxies. 

As a second type of example we mention a problem that is amenable 
to considerable mathematical analysis as well as to computational attack. 
This is the question of the spiral structure of galaxies. The so-called 
density wave theory approaches this question as a dynamical problem in 
the form of the gravitational instability of a galactic dish with respect to 
spiral modes. There are three basic approaches hi the calculation of spiral 
structures of galaxies on the basis of density wave theory that may be 
characterized loosely as the stellar model, particle model, and fluid model. 

The stellar model represents the classical approach in the study of 
stellar systems. In brief, a galaxy has a stellar component and a gaseous 
component. The latter usually has a sufficiently small mass to be negligible 
in first approximation. The basic equations governing the behavior of the 
stellar component are the Boltzmann equation, usually in colli sionless form 
when close encounters between stars are omitted, and the Poisson equation 
for the gravitational potential. In a sense, these equations are much simpler 
than the Vlasov equations of plasma physics, but their numerical solution 
nevertheless poses challenging questions. 

In the particle model, the stellar component is considered as a very 
large but finite number of particles. As in our first example, the computa- 
tional approach then assumes the form of a simulation process. In brief, the 
motions of the individual stars are followed, and their gravitational field 
is calculated by a self-consistent evaluation of the field. This technique 
has provided valuable qualitative information. But the number of stars in 
these models is usually of the order of 10 11 - 1 , and hence, by necessity, 
numerical simulations are extremely limited in accuracy and provide only 
few quantitative data for specified galaxy models. 

In the third approach, the stellar system is considered to be a con- 
tinuous medium. In this setting one may then study the characteristics 
of wave patterns over the galactic disk and then dependence on the mass 
distribution. In particular, the spiral structure of galaxies may be ex- 
plained in terms of spiral wave patterns of some kind. Such stationary 
wave patterns over a field of differential rotation are to be expected; in 
fact, hydrodynamic waves over shear flows are known to exist for some 
time especially in the form of self-excited modes. The general form of 
the resulting theory of spiral galactic structures requires elaborate com- 
putational techniques. At the same tune, asymptotic approaches have 
'provided analytic results that have led to a better understanding of the 
dynamical mechanisms despite their limitations in accuracy and in the 
types of galaxies covered. 

As noted earlier, these are only some examples of the many different 

36 



types of problems in astrophysics that relate closely with applied and com- 
putational mathematics. Many of these problems involve extremely wide 
ranges in length and time scales. For example, calculations of stellar evolu- 
tions have to range over billions of years, while the dynamics, say, of the 
supernovae phase takes place within milliseconds. Similarly, the size of a 
neutron star differs by substantial orders of magnitude from the size of its 
corresponding relevant gravity field. Thus, in particular, the mathematical 
and computational difficulties, discussed in Section 3.2, are especially ap- 
plicable here. Moreover, the wide range of the underlying physical theories 
leads to many substantially different types of mathematical models, which 
in turn require very different computational techniques. 



2.9 STRUCTURAL MECHANICS 

During the past two decades, the use of computers has transformed 
large parts of solid mechanics into practical tools for a multitude of 
technological developments. Sophisticated computational software is 
employed throughout the nation's industries and research laboratories in 
the analysis and design of structures and mechanical equipment. There 
is a strong interaction between applied mathematics and solid mechanics. 
Mathematical analysis has provided insight into model formulation and 
the development of powerful numerical methods; and, vice versa, novel 
engineering approaches have led to new research areas in applied mathe- 
matics. 

In the case of linear problems there exists now a relatively broad 
experience in computing solutions for a range of problems concerning 
the behavior of solid bodies subjected to specified loads. In general, 
the computed results are considered reasonably reliable, and they have 
been corroborated over a period of time by observation and practical 
experimentation. There appears, however, to be a growing need for the 
development of computable error estimates that can provide a realistic 
check on the solution accuracy. Such a posteriori bounds have been shown 
to be feasible; but their applications to large, realistic problems in solid 
mechanics still requires considerable research and software development. 
The general availability of economical a posteriori estimates would make 
possible the consistent use of adaptive mesh-refinement techniques that 
would reduce the cost of data preparation by the users and make it possible 
to generate near-optimal solutions for a given amount of computational 
expense. 

37 



The general trend in computational solid mechanics today is toward 
extending the computational methodology to nonlinear problems. Sources 
of nonlinearity in structural problems are (1) geometric nonlinearity due 
to nonlinear strain-displacement relations, (2) material nonlinearity due 
to nonlinear constitutive equations, (3) force nonlinearity due to nonlinear 
stress boundary conditions, and (4) kinematic constraint nonlinearity due 
to nonlinear displacement boundary conditions. The source of nonlinearity 
affects the form of the resulting nonlinear equations and, hence, influences 
the effectiveness of the solution techniques. 

The numerical analysis of all of these nonlinear problems is not yet 
at a satisfactory stage. Many computer programs for such problems exist, 
but the mathematical basis for most of the methods used is insufficiently 
understood, and there is little known about the accuracy of the computed 
results. Moreover, in the case of nonlinear problems, few numerical com- 
putations can be supplemented with sufficient experimental experience. 

The situation is still best understood in the case of finite elasticity. 
Even there the mathematical theory of the underlying equations is incom- 
plete, and the approximation theory for these equations is generally based 
on various simplifying assumptions that may or may not be valid for a 
particular problem. 

The state of the art in elasto-dynamic problems is in even worse 
shape. It is known that multiple solutions may exist and shock waves can 
develop. Moreover, not all solutions are necessarily physically relevant. 
The questions of how to model such phenomena numerically and how to 
determine the physically realistic solutions are as yet largely open. When 
it comes to problems in finite plasticity even less is known. Although there 
has been much progress in this area during recent years, no satisfactory 
and complete mathematical model is available as yet. Especially, there 
are profound mathematical and computational difficulties in modeling 
phase changes, viscous effects, cracks and singularities, the growth of 
cracks under dynamic loading, and the identification and implementation 
of physically reasonable constitutive equations describing these materials. 



2.10 NONDESTRUCTIVE TESTING AND TOMOGRAPHIC 
RECONSTRUCTION 

There is a pervasive need in technology to evaluate quantitatively the 
integrity and the remaining reliable lifetime of components and structures, 
e.g., from bridge girders to high-performance ceramic turbine disks. In 

38 



the past decade considerable progress has been made. This developing 
technology is called nondestructive evaluation (NDE) to distinguish it 
from older nonquantitative nondestructive testing practices. NDE presents 
challenges to the applied mathematical sciences on many levels. A few of 
the more important and topical problem areas are mentioned here. 

In NDE applications a component is often subjected to some sort of 
penetrating radiation with the aim of deducing information about its in- 
ternal state from a measurement of the radiation field external to the part. 
Examples include the use of ultrasonic radiation, x rays, and neutrons. 
Because of its flexibility, relative cost, and safety, ultrasonic methods are 
often used in NDE applications. The deduction of information about flaws 
from the incident and scattered ultrasonic fields relies on the solution or 
approximate solution of the inverse scattering problem for elastic waves. 
In some regimes it is possible to develop and adapt imaging techniques 
to the ultrasonic setting. In either case information about defects must 
be obtained from band-limited noisy data. A fundamental limitation to 
our current ability to utilize ultrasonic techniques broadly is our limited 
understanding of the elastic inverse scattering problem in either the fre- 
quency or the time domain. 

In other important NDE applications, electrical currents are induced 
in a material. These currents produce fields that vary depending on 
whether a defect, e.g., a crack, is present in the material. The utilization 
of these so-called eddy-current techniques depends on the ability to infer 
information about the defects from the measured fields they produce. This 
again is an inverse problem that is imperfectly understood. 

Passive methods in NDS are also widely used. For example, when a 
pressure vessel or aircraft is in service, a crack once formed may grow and 
propagate. This will be accompanied by the release of acoustic energy. 
This energy can in turn be monitored at selected sites. The problem then 
becomes one of identifying and classifying the sources of these sound pat- 
terns using acoustic emission studies, an increasingly important technique 
in NDE. In mathematical terms the problem is the so-called inverse source 
problem, which is beset with the same type of difficulties that the inverse 
scattering problems possesses. Sparse, noisy data often taken at highly 
nonoptimal locations are the raw information from which source charac- 
teristics must be deduced. 

The techniques of NDE have application in other areas, and much 
can be learned in other applications fields that is valuable to NDE. For 
example, there is a close connection between the need for inverse scattering 
results in geophysics and NDE and acoustic imaging results in NDE and 
biomedical applications. An area of great success in medical applications, 
tomography, can provide useful information in selected NDE applications. 

39 



The greatest success of tomography has been in medical applications 
such as the CAT (computer assisted tomography) scanner. Unlike ordi- 
nary x-ray technique, which masks important features by superposing on 
a single picture information on planes perpendicular to the beam, com- 
puterized x-ray tomography provides pictures of individual thin slices 
through the body. Several hundred parallel x-ray pencil beams are 
projected in the plane of the slice, and the attenuation of the x rays in 
each beam is measured separately and recorded. The procedure is repeated 
for many different beam directions. An elaborate calculation then permits 
approximate reconstruction of the x-ray attenuation density as a function 
of position within the slice. 

The idealized mathematical problem is the reconstruction of a func- 
tion of two variables from its integrals along lines. This problem, as well 
as its three-dimensional version, was solved by Radon in 1917 and later 
rediscovered in various settings such as probability theory (recovering a 
probability distribution from its marginal distributions) and astronomy 
(determining the velocity distribution of stars from the distribution of 
radial velocities in various directions). Of course, much work was needed 
to adapt the Radon inversion formula to the incomplete information avail- 
able in practice. Various algorithms for the numerical inversion of this 
ill-posed problem have been proposed, with the present trend favoring the 
so-called convolution algorithm on account of its speed and accuracy. Each 
area of application has, however, its own requirements and may need a 
modification of existing reconstruction algorithms or even a custom-made 
one. Some algebraic methods, for instance, can easily incorporate a priori 
information about the object to be reconstructed. 

Recent advances in medical tomography include nuclear magnetic 
resonance (NMR) tomography and positron emission tomography (PET). 
In NMR strong magnetic fields are used to affect the nuclear magnetic 
spin rate of hydrogen atoms. By varying the fields and their direction, the 
plane integrals of the density of hydrogen can be measured and the den- 
sity reconstructed by an algorithm based on the above three-dimensional 
version of Radon's theorem. The technique is now regarded as competi- 
tive with x-ray tomography for many purposes; and, of course, it is not 
ionizing The advantage of PET over CAT is that metabolic processes 
can be followed. A compound such as glucose is made using carbon-11 
atoms which emit positrons. Photons resulting from the annihilation of 
an emitted positron with an electron are detected by a bank of detectors 
that can record coincidences. Recently, algorithms based on probabilistic 
arguments have been proposed for the PET reconstruction problem 



40 



2.11 MATHEMATICAL MODELS IN THE BIOLOGICAL SCIENCES 

The biological manifestations of the physical laws of the universe present 
us with a rich variety of new phenomena that require the devlopment 
of new mathematical tools and computational methods. We shall discuss 
just a few examples of mathematical research in cardiovascular physiology 
and neurophysiology, with the knowledge that there are many other areas 
of biological sciences in which mathematics and computing are fruitfully 
applied. 

Blood flow in the heart obeys the incompressible Navier-Stokes equa- 
tions, which, in turn, are simply a statement of Newton's laws in 
differential form (see Section 2.1). The distinctive biological character 
of the problem comes, however, from the moving boundaries that are in 
contact with the blood. These include the muscular heart walls and the 



CURVED PIVOTING DISC VALVE EXP.4: 1-9 



STREAMLINES 





\ 





FIGURE 2.10 Computer-generated plots showing the predicted opening 
movement of a curved pivoting-disk valve mounted in the mitral (inflow) 
position of the left ventricle. The curvature makes the valve open more 
widely than a straight valve pivoted at the same point. It also helps 
to prevent stagnation in the smaller opening of the valve. [Prom D. M, 
McQueen and C. S. Peskin, Computer-assisted design of pivoting-disc 
prosthetic mitral valves, J. Thorac. Cardiovaac, Surg. (in press),] 



41 




1.4 

FIGURE 2.11 Transmission and "reflection" of a nerve impulse at a junc- 
tion where the diameter of the neuron suddenly increases. The plots show 
computed voltage as a function of tune at equally spaced positions. The 
junction is at x = XQ, and the ratio of diameters there is 2.5:1. Note the 
increase in propagation speed for x > XQ. A reflected wave is set up when 
the larger fiber re-excites the smaller fiber after the refractory period of the 
smaller fiber has elapsed. [From S. S. Goldstein and W. Rail, Changes of 
action potential shape and velocity for changing core conductor geometry, 
Biophysical J. 14, 731 (1974).] 



elastic heart valve leaflets. The motions of these boundaries are not known 
in advance; they must be computed along with the motion of the fluid. 

These considerations have led to the development of a computational 
model of the left heart that can be used in the computer-aided design 
of artificial heart valves. In this model, the fluid equations are solved 
by finite-difference methods on a regular, square mesh (see Section 3.4). 
The boundaries are represented in Langrangian form as a collection of 
moving points. Coupling coefficients between boundary markers and fluid 
mesh points are computed with the aid of an approximation of Dirac's 
^-function. This computer model (Fig. 2.10) has been used in the design 
of prosthetic heart valves to remedy problems of stagnation and blood 
clotting in the smaller opening of the valve. The model has also been 
helpful in the study of disease processes, providing, for example, a possible 
explanation for mitral valve prolapse. 

42 



Just as the flow of blood in the heart is ultimately governed by 
Newton's laws, the conduction of electrical signals along nerves is ul- 
timately governed by Maxwell's equations (Figs. 2.11 and 2.12). Here 
also, nature provides a peculiar boundary condition that leads to entirely 
new phenomena. In this case the boundary condition comes from certain 
voltage-dependent channels located in the nerve membrane. These intro- 
duce a nonlinearity, and the equations of nerve conduction take the form 
of nonlinear diffusion equations the Hodgkin-Huxley equations. Without 




FIGURE 2.12 Computed electrical signals at the input end of a neuron. A 
brief pulse of current is applied at the periphery of a tree, and the result- 
ing voltages are computed (logarithmic scale) at the input (BI), at succes- 
sive branch points (P, GP, GGP), and finally at the cell body (SOMA). 
[Modified from J. Rinzel and W. Rail, Transient response in a dendritic 
neuron model for current injected at one branch, Biophysical J. 14, 759 
(1974).] 

43 



the nonlinearity, signals introduced at one end of the nerve would decay 
rapidly; because of the nonlinearity such signals evolve into a specific 
waveform that propagates at a constant speed without distortion. This 
waveform is called the nerve impulse, and it is the basic unit of long- 
distance communication in the nervous system. 

Currently, there is a wide range of mathematical, computational, and 
physiological research activity related to the Hodgkin-Huxley equations. 
Mathematically, there is extensive research on the basic theory of non- 
linear diffusion equations. A particularly fruitful approach here has been 
the use of piecewise linear models that expose the basic structure of the 
equations. Singular perturbation methods have also been useful because 
the equations exhibit a disparity of time scales. 

An important physiological enterprise is the modification and applica- 
tion of the Hodgkin-Huxley equations to other excitable tissues. In the 
heart, for example, equations of the Hodgkin-Huxley type describe the 
electrical processes that generate the cardiac rhythm and coordinate the 
heartbeat. Mathematicians are just beginning to use these equations as 
a basis for a theory of the abnormal rhythms of the heart, of which the 
most serious is ventricular fibrilation. This theory has connections with 
recent work on chaotic dynamical systems and transition to turbulence: it 
appears that fibrilation is directly analogous to turbulence. This work has 
enormous practical significance, since the principal cause of sudden death 
following heart attacks is ventricular fibrilation. 

Progress has also been made in the modeling of the input to the 
neuron (whose output signal is the nerve impulse). The neuron integrates 
information received through a tree of dendrites in which the signal- 
ing mechanism is often described by the linear diffiision equation with 
leakage. Mathematical modeling of the dendritic tree has had a substan- 
tial impact on experimental neurophysiology. One reason for this is that 
dendrites are too small to be penetrated with microelectrodes. Thus the 
neurophysiologist can only record voltage or inject current at the cell body 
and is forced to rely on the theory to indicate the significance of these 
measurements with respect to activity in the dendritic tree. 

Some major successes of the theory are as follows: 

1. Elucidation of the dramatic differences between effects of a synapse 
close to the cell body and effects of a similar synapse far out in the dendritic 
tree. 

2. A possible explanation of the role of dendritic spines in learning 
and memory. 

3. Prediction of the existence of dendro-dendritic synapses based 

44 



on a mathematical model of field potentials in the olfactory bulb. Such 
synapses, previously unheard of, were subsequently found in electron 
micrographs. This work led to a fundamental new concept of local in- 
formation processing in an "analog" rather than a "digital" mode, i.e., 
without nerve impulses. Such processing is important in neural networks 
such as the retina. Indeed, an extensive mathematical theory of the retina 
has been developed, and this is another exciting area of current research. 



2.12 ELECTRONIC COMPONENTS 

The design and fabrication of modern integrated circuits is a complex 
process. The number of devices that one can put on a chip depends on 
the size of the chip and how small one can make its features. Over the 
years the largest increase in the number of devices on a chip has resulted 
from the continuing reduction in feature size and with this a reduction in 
device size. 

Consequently, process and design engineers have had continually to 
redesign the process steps and then recalculate the resulting device charac- 
teristics to ensure good electrical behavior. This has had to be computa- 
tional because the equations are mathematically intractable and a trial and 
error approach is prohibitively expensive and tune-consuming. Moreover, 
experimental techniques tell us only what happened, not why. Effective 
device design depends on determmmg both the what and the why by vary- 
ing the problem parameters in the computational model. 

The mathematical models on which the theory of semiconductor 
devices rests are differential equations that describe the flow of current 
(holes and electrons) under the influence of electric fields. When feature 
sizes were large, the devices could be treated as though they consisted 
of plane surfaces and edge effects could be neglected. This allowed the 
development and successful use of one-dimensional analytical models ob- 
tained by solving a system of three coupled, nonlinear ordinary differential 
equations. As device sizes shrunk, these models became more complicated 
and less accurate as edge effects became more important. In very large- 
scale integration where device dimensions have reached a few micrometers, 
these models are no longer adequate, and the coupled, nonlinear, partial 
differential equations must now be solved in two and sometimes three 
dimensions. These differential equations consist of a nonlinear Poisson 
equation that describes the potential of the electric field and two non- 
linear transport equations that describe the motion of the holes and the 

45 



electrons. Moreover, the smaller sizes have made some physical effects 
important whereas they were previously neglected. This results in further 
complications to the basic equations. 

Partial differential equations play an increasingly important role 
in simulating the fabrication processes. The transistors in a chip are 
formed by implanting certain dopant ions into selected areas of the chip. 
Subsequent high-temperature processes, such as growing an oxide layer, 
will cause these atomic impurities to diffuse. Their final distribution is an 
important factor in determining device characteristics. Insights into these 
fabrication processes are especially important for increasing the yield of 
reliable devices, which is a critical factor in their economic viability. 

The design of the overall circuit to be placed on a chip leads to large 
systems of nonlinear differential equations that need to be solved numeri- 
cally. Then the efficient layout of the circuit on the chip introduces com- 
binational and graph theoretical problems, which again pose formidable 
computational problems. 



46 



3. COMPUTATIONAL AND MATHEMATICAL DIFFICULTIES 



Computational mathematics and modeling is a relatively young science, 
one that is expanding rapidly. The success and importance of the field 
stems from the fact that its application provides the possibility of tack- 
ling significantly more complex and difficult problems than would other- 
wise be possible. Perhaps the greatest opportunity provided by a com- 
putational approach is that it opens the wide realm of strongly non- 
linear phenomena to systematic, relatively accurate, and efficient model- 
ing, improving the chance that important phenomena can be isolated 
and analyzed. Nonlinearities pervade nearly all aspects of applied mathe- 
matics, and to a large extent these nonlinearities are responsible for the 
difficulties that are encountered in computational modeling. Our purpose 
in this chapter is to explore the source of some of these difficulties. In 
Chapter 4, we will discuss the direction of some of the computational 
research needed to resolve the problems. 



3.1 DEGREES OF FREEDOM 

There are several reasons for the degrees of freedom in a model, and 
hence the size of computational problems, to increase. One is that we 
attempt to increase the accuracy and complexity of our representations of 
the physical conservation laws. Increased degrees of freedom come either 
from increasing the number of dependent or independent variables. For ex- 
ample, in chemical models the number of dependent variables is increased 
by increasing the number of chemical species considered. An obvious 
need for increased independent variables comes from the need to represent 
phenomena in two and three spatial dimensions. However, even higher- 
dimensional problems arise when the independent variables are not the spa- 
tial coordinates but are various state descriptors; such higher-dimensional 
problems are common in physics and chemistry, Unfortunately, the direct 
application of numerical methods that work well in one or two dimensions 
often are not usable in three dimensions. Therefore, increasing the degrees 
of freedom may require significantly different algorithms. 

47 



Another way to increase the degrees of freedom in problems is to 
allow the possibility of multivalued solutions. Certain physical systems 
allow the solutions to bifurcate, a phenomenon discussed in Section 3.5. 
Other systems have hysteresis or "history-dependent" properties. In these 
problems the solutions depend not only on the boundary conditions but 
also on the path that the transient solution follows. Structural dynamics 
models for plastic materials in which the stress depends not only on the 
strain but on the straining rate as well exhibit this behavior. 

Consider an example to quantify the magnitude of the computational 
requirements for multidimensional problems. First suppose we want to fol- 
low the evolution of a chemical model having N chemical species. Suppose 
also that the problem is stiff (see Section 3.2) so for each time step in 
the evolving system we solve N nonlinear algebraic equations in N un- 
knowns. .Even for models with a large number of species this is a relatively 
straightforward task. However, if we now want to introduce transport 
phenomena, such as fluid mixing, the problem has to include the spatial 
dependence of each chemical species. 

For a one-dimensional case assume that we use I spatial mesh points 
and that we estimate what each species will do on the basis of its current 
local value and that of its immediate neighbors (a three-point spatial stencil 
involving all species). We now have NI unknowns and a nonlinear system 
that has 3NI nonzero entries in each equation. It is typical to have 40 to 50 
species and to require 100 mesh points to resolve the species concentrations 
accurately. In this case over three quarters of a million words of memory 
are needed just to store the approximating local linear system along with 
the solution. This alone is larger than the fast-access memory of most 
modern computers. Suppose further that the same model is to be posed 
in two dimensions on a I x I mesh and in three dimensions on an / X / X / 
mesh. The solutions themselves require NI 2 and NI 3 words of storage, 
respectively. Worse, the totality of coefficients in the typical linear system 
will be 5N 2 / 2 and 7N 2 I 3 . It is of practical interest to want solutions to 
three-dimensional combined kinetics and transport problems. However, 
even for a modest system of 20 species and 50 mesh points per spatial 
dimension, 2.5 million words are needed to store the solution alone. An 
additional 350 million are required to store the coefficients of the linear 
system. (For comparison, we note that the largest computers currently 
available have 4 million words of fast-access memory.) Moreover, in the 
high-dimensional cases, the linear system is not conveniently structured 
for efficient solution. 

Clearly a major problem in solving high-dimensional systems of par- 
tial differential equations is that after discretization the resulting system 
of approximating linear equations can be much too large to be solved 

48 



effectively by direct means. Except in simple cases the resulting equations 
have to be solved iteratively, and, especially for strongly nonlinear equa- 
tions, the convergence properties of the iterative process may be a major 
concern. 

Operator splitting methods, such as alternating direction implicit 
(ADI) (see Section 4.7), also provide a current effective approach for 
high-dimensional problems. The approach here is to alternate the solu- 
tion of a set of lower-dimensional problems. The alternated problem ap- 
proximates the original problem with sufficient accuracy, but the set of 
lower-dimensional problems is much more easily solved, even in aggregate. 
Again convergence and accuracy properties must be established in all but 
the simplest of cases. 

For some problems it may be more efficient to depart from the con- 
ventional ideas of discretization on a mesh network and consider instead 
mesh-free methods or (see Sections 4.8 and 4.9) Monte Carlo methods. 
Unknowns can be changed: instead of the amount of material at a given 
location, one can ask for the amount of a given wavelength in the solution 
as a whole. In fluid mechanics, vortex methods could be a more efficient 
approach. Such methods reduce the size of the linear algebra problems, in 
comparison with the mesh-oriented methods, if the new unknowns carry 
all the important information desired. 

Decisions about which methods are most effective may depend 
strongly on advances and changes in computer architecture. Since new 
architectures, such as vector and parallel processors, are now evolving, 
the numerical analyst has to re-evaluate his approaches periodically. Also, 
considerations such as the relative cost of memory versus central processor 
time can bear heavily on decisions regarding algorithms. 



3.2 DIFFERENT TIME AND LENGTH SCALES 

In principle a model can contain important length scales that range from 
the size of an atom to the size of the universe. In practice, however, we 
limit the range of scales by approximation (see, for example, Section 3.7). 
Nevertheless, it is often the case that mathematical models of physical 
processes are characterized by the simultaneous presence in their solution 
of significantly different time and length scales. The solutions to these 
models will have regions of strongly localized behavior, such as shocks, 
steep fronts, or other near discontinuities. Therefore important topics of 
research in numerical analysis are the consideration of such circumstances 

49 



and the development of efficient theories and methodologies for their com- 
putational solution. Indeed, we often find situations for which solutions 
are not possible, or at least not practical, without the application of spe- 
cialized methods to deal with the multiple characteristic scales. 

Fluid mechanics and chemical kinetics are two areas that provide a 
rich source of examples for multiple and disparate scales. Fluid-mechanical 
processes are commonly characterized by groups of parameters that are 
indicative of the various scales in a problem. Some examples are the Mach 
number (relates velocities to sound speed), the Reynolds number (relates 
inertia! forces to viscous forces), the Prandtl number (relates viscous effects 
to thermal effects), and the Damkb'hler number (relates chemical reaction 
rates to diffusion rates). When any of these numbers is very large (or 
small) it is likely that the solutions to the models will have regions of 
localized behavior. For the case of large Mach number the possibility of 
shocks exists. Similarly for large Reynolds number we expect to encounter 
boundary layers in the vicinity of solid walls. When the Prandtl number 
is large we expect that thermal boundary layers will be much thinner 
than viscous boundary layers. And when Damkohler numbers are large we 
expect narrow reaction fronts. 

In chemical kinetics we find large numbers of chemical reactions com- 
peting for the available chemical species at widely different rates. As a 
result some species are either consumed or produced rapidly or slowly, 
while other species are being both produced and consumed simultaneously 
at high rates, with their net production rate being relatively slow. This 
chemical behavior is responsible for the many widely differing time scales 
in the mathematical models. The computational models of these processes 
are characterized as either multirate problems or stiff problems. 

It is worthwhile to point out the distinction between multirate prob- 
lems and stiff problems. In both problems the system itself is equally 
capable of rapid or slow changes. Multirate problems are those in which 
at least one component of the solution is changing rapidly, even though 
others are changing slowly. Numerical methods for these problems must 
take time steps that are small enough to resolve the fast transients, so they 
are controlled by accuracy not stability considerations. Stiff problems, on 
the other hand, are those in which all components of the solution axe 
changing slowly compared with the fastest characteristic scales possible in 
the model. In these cases explicit numerical methods are forced to take 
much smaller tune steps than are needed to maintain accuracy in order 
tomamtam stability. Often problems that begin as multirate problems 
become stiff problems as an equilibrium or steady-state condition is ap- 
proached. Stiff problems are usually solved efficiently by implicit methods. 



50 



For stiff or multirate problems, it is perhaps useful to consider methods 
that treat the fast components differently than the slow components. 

Given that multiple-scale problems are of practical importance, we 
must consider: why are there computational difficulties, and what can be 
done to ameliorate the difficulties? In the case of disparate length scales 
one difficulty is that of representing accurately the highly localized be- 
havior. If the solution is represented discretely on a mesh network, the 
mesh must be sufficiently fine to capture the localized behavior accurately. 
The whole topic of adaptive meshing is critically important for these prob- 
lems (see Section 4.2). Here, instead of computing on a fixed prespecified 
mesh, the mesh adjusts itself dynamically as the solution develops hi order 
to maintain accuracy in the solution. 

For situations in which the localized behavior is known to be ap- 
proximated well by very sharp fronts (e.g., shocks or flames), front track- 
ing methods can have significant advantages. Unlike the adaptive mesh 
approach where the solution is resolved smoothly through the front, the 
front tracking methods approximate the front by a discontinuity whose 
magnitude, speed, and location are to be found. Then elsewhere in the 
region the conventional discrete representations are adequate. 

In problems like chemical kinetics, the disparate time scales cause 
the governing differential equations to be stiff. Here, explicit solution 
methods are well known to be extremely inefficient, and some form of 
implicit method is needed. For systems of ordinary differential equations 
the problem has been worked out, and high-quality computer software is 
available. However, when stiffness is encountered in the context of systems 
of partial differential equations the remedies are much less developed. The 
same techniques used for ordinary differential equations, when applied 
directly the partial differential equation problems of practical interest, 
often yield problems that are simply too large for current computers. 

Several approaches show promise. One is to develop operator- splitting 
methods in which the stiff parts of the problem are split off and solved as 
a series of smaller and hence more tractable problems. Another approach 
is to attempt to remove the stiffness by solving instead an approximate 
(yet sufficiently accurate) system of equations. This tack benefits from an 
asymptotic analysis of the equations. Both approaches have found recent 
successes in fluid mechanics and in combustion chemistry. 



51 



3.3 SINGULARITIES IN COEFFICIENTS, DATA, OR STATES 

Difficulties similar to those encountered in multiple-scale models are often 
found in problems having singularities in coefficients or states. That is, a 
singular or discontinuous coefficient can give rise to localized behavior in 
the solution, such as very steep fronts. An example could be a material 
interface in a structural or heat-transfer problem, say between a steel and a 
plastic part. At this material interface the solution (stress or temperature 
gradient) might change rapidly. In order to maintain accuracy in the 
computed solution, the numerical procedure would have to resolve this 
frontlike behavior. The situation is analogous to the occurrence of a shock 
or a flame front. However, we usually know where the material interfaces 
are, so they are perhaps easier to deal with than phenomena such as shocks. 
Phase transitions can also produce discontinuous coefficients. Take 
a model in which a melting front is traveling through a region. Usually 
the properties of the molten material are quite different than those of the 
solid material. In fact, different governing equations may even be required 
for the two regions (e.g., fluid motion may be modeled in the liquid but 
not the solid). In any case, the solution is likely to exhibit jumps in its 
properties at the transition, and the numerical method will need to locate 
and resolve it. This situation is more like a shock, in that the position of 
the phase transition front is not known a priori, and thus the numerical 
method must both locate and resolve it. 



3.4 BOUNDARY CONDITIONS 

The solution to a boundary or initial-boundary-value problem depends 
strongly on the boundary conditions. Thus it is important to understand 
the relationship of the boundary conditions to the differential equations 
and to their discrete representation. Most important, the boundary con- 
ditions must be chosen so that the problem is well posed. For a large class 
of problems there is a satisfactory theory of admissible boundary condi- 
tions, but for many problems, those involving coupled hyperbolic-elliptic 
systems or disparate time scales, for example, only a rudimentary theory 
is available. 

A possible error in prescribing boundary conditions for hyperbolic 
equations is to overspecify or underspecify the number of boundary condi- 
tions. Overspecification usually causes nonsmooth solutions with mesh os- 

52 



dilations near the boundary. Underspecification does not ensure a unique 
solution, and the numerical solution may tend to wander in steady-state 
calculations. In either case, the results are not accurate and one should 
be skeptical of even the qualitative behavior of the solution. It should 
be noted that the way in which boundary conditions are specified for the 
difference equations can change a well-posed continuous problem into an 
ill-posed (unstable) discrete problem. 

Two of the most common methods used to incorporate boundary 
conditions into discrete equations are the extrapolation and uncentered 
differences methods. In the extrapolation method, the domain of the prob- 
lem is extended and the solution is extrapolated to fictitious points outside 
the integration region. The nonphysical solution at these points is defined 
so that the discrete equations are consistent with as many relationships 
as can be derived from the boundary conditions and differential equa- 
tions. The extrapolation formula can do this best by incorporating the 
discrete boundary conditions into the extrapolant. Additional relations 
can be generated by differentiating the boundary conditions with respect to 
time, replacing all time derivatives by space derivatives using the original 
differential equation, and discretizing the resulting equations. 

The uncentered differences approach is to extend the number of 
boundary conditions so that all components of the solution are defined 
at the boundary. Again, these additional boundary conditions must be 
consistent with the original problem and as many relationships as can be 
derived from it. An uncentered difference approximation is then used to 
approximate the spatial derivatives at the mesh points nearest the bound- 
ary. 

Irregular domains can be imbedded in an underlying regular grid 
that is not aligned with the boundary, or an irregular grid can be con- 
structed that conforms to the boundary. The discrete approximations to 
the equations away from the boundary are much simpler on the regular 
imbedded grids, but the boundary conditions are difficult to approximate. 
Boundary-fitted grids can be generated algebraically in the original physi- 
cal domain, or the domain (and hence the grid) can be mapped onto a 
regular grid in a simpler domain and the equations solved there. The 
algebraic-grid-generation methods have the advantage that the equations 
and boundary conditions are unchanged, but the differential operators are 
more difficult to approximate on the nonuniform grid. When using the 
mapping method, the differential operations are easily approximated but 
the transformation can greatly complicate the equation and sometimes 
obscure important properties such as the conservation laws expressed by 
the equations. 



53 



3.4.1 Boundary Conditions at Infinity 

Many physical problems require the solution of partial differential equar 
tions on some infinitely large domain fL For computational reasons this 
domain is often replaced by a finite domain QI. Then the difficult problem 
of specifying boundary conditions at its finite artificial boundary B arises. 
It is especially important that these artificial boundary conditions do not 
introduce spurious phenomena. Consider, for example, a nonviscous fluid 
that at subsonic speed leaves HI through the boundary B. There is one 
characteristic direction that points back into the region HI, and there- 
fore one boundary condition has to be specified on B. But, in general, 
no detailed knowledge of the solution on B is known and therefore other 
principles have to be applied. For example, if one has solved the problem 
by difference approximation then one predicts the solution on B from in- 
side wi, for all the dependent variables. Thus this procedure amounts to 
overspecification of the solution on B. Another principle has been proposed, 
namely, to specify the boundary conditions on B so that no reflection of 
high frequency takes place. However, numerical experiments have shown 
that such approaches do not always work. 



3.5 BIFURCATIONS AND CHAOS 



3.5.1 Buckling and Collapse Behavior, Bifurcations 

In general, the equilibrium equations of a mechanical structure involve a 
finite number of parameters, that is, they have the generic form F(x, p) = 
0, where x varies in some state space X and p 6 R m represents a parameter 
vector. Thus, in general, the solution set {(x,p) 6 X xR m ; F(x,p) = 0} is 
a manifold X x/?" 1 , and one topic of interest is the location and character 
of the singular points of the solution set. 

Without going into detail, suppose that we follow some curve on 
the solution manifold defined by some combination of parameter values 
with one degree of freedom represented by a single variable X. Then 
we may encounter certain critical points on the solution curve where 
the mechanical structure may suffer a loss of stability. Such a loss of 
stability actually corresponds to a dynamic phenomenon whereby the 
structure undergoes a sudden change of deformation. The dynamics of this 

54 



phenomenon are not described by the equations of equilibrium F(x, p) = 0, 
but it is possible to deduce from the shape of the (static) solution manifold 
at that point what type of sudden changes may be expected. 

We use a few figures to illustrate the situation. In the upper left-hand 
part of Fig. 3.1 the point denoted by (1) is a so-called limit point or turning 
point and an increase of the load-intensity X beyond the critical value X c 
results in a jump from (1) to (presumably) (2). This type of behavior is 
called a snap-through or collapse. In the case of the upper right-hand part 
of Fig. 3.1 the instability phenomenon is related to the bifurcation of the 
solutions at (1). This behavior is called buckling. This part of the figure 
is a classical example of stable buckling where a distinct change in the 
character of the load deformation is encountered when the load-intensity 
X passes the buckling load X c but where no failure of the structure occurs. 
On the other hand, the lower part of Fig. 3.1 shows an unstable buckling 
point where we again observe a dynamic departure from the bifurcation 
point to some other state [presumably the equilibrium state (2)]. The 
geometrical shape of the bifurcation branch n is the determining factor in 
the question of whether the bifurcation point is stable or not. 

These examples already indicate that for a deeper understanding of 
the behavior of a mechanical system it is necessary to analyze the shape of 
the full solution manifold. Of course, the choice of the parameters entering 
into the definition of the equation is of critical importance here. In essence, 
catastrophe theory provides some information about the selection of par- 
ticular minimal numbers of parameters, but in practice the parameters 
are simply chosen to correspond to the certain natural features of the 
structural problems. 

In view of these observations the aim is to develop procedures for 
a computational analysis of the form of the equilibrium surface. Some 
methods for this purpose are mentioned in Section 4.4 on continuation 
methods, but the entire problem is still a wide open research problem. In 
particular, it has to be noted that we can compute only points that are 
approximately on the solution manifold of a discretization of the original 
problem. Thus, the questions arise whether phenomena, such as limit 
points, or stable or unstable bifurcation points, encountered on the solution 
set of the discretized problem actually correspond to similar phenomena 
for the original problem, and, if so, what errors have been encountered. 
These questions are as yet largely unanswered and represent considerable 
research challenges. In this connection, it might be mentioned that for 
nonlinear problems of this type the solution manifold of the discretized 
equations often has a different number of connected components than 
that of the original equations. The components that do not approximate 
the exact solution manifold have been called spurious, or numerically 

55 



F=XF 





dliplacamant W 



cylindrical shell - roof segment 

Collapse Behavior 

F=XFo 




dliplooamant W 



simply supported plate 

Stable Buckling 



F=XF 




dliplooamant W 



cylindrical shell segment 
In compression 



Unstable Buckling 

FIGURE 3.1 Instability behavior of shells under different loadings F. On 
the right the load intensity X is plotted versus a characteristic deflection 
W. [Prom W. Rheinboldt and E. Riks, "A survey of solid techniques 
for nonlinear finite element equations," State of the Art Surveys of Finite 
Element Technology, Chap. 16, special AMD pub!., American Society of 
Mechanical Engineers, New York (1983).] 

56 



irrelevant solutions. They are being observed more and more often in 
various applications, but their study has only recently begun. 



3.5.2 Chaos in Deterministic Systems 

It has been remarked that one of the most engaging problems in nonlinear 
dynamics is that of understanding how simple deterministic equations can 
yield apparently random solutions. As is now widely recognized, systems 
with this behavior appear hi many of the branches of scientific endeavor. 
This near universality has come to the attention of physical scientists and 
applied mathematicians principally through the use of computers in the 
study of properties of dynamical systems. 

The onset of chaos in deterministic systems (the stochastic instability) 
signals an unusual sensitivity to initial conditions, i.e., the trajectories hi 
phase space diverge widely as tune goes on even though the initial con- 
ditions are arbitrarily close to one another. The behavior is such as one 
would expect hi a space with everywhere negative Gaussian curvature. On 
the other hand the trajectories may hi fact tend to a single orbit nearly 
filling a subspace of the phase space (a strange attractor) in an ergodic 
manner. In even the simplest systems of this type it is possible for several 
such attractors to coexist side by side, with the initial conditions deter- 
mining which one is reached asymptotically hi tune. These and other 
numerous unusual properties of dynamical systems displaying chaotic be- 
havior have led to some open problems in computational mathematics. 

Examples of such problems include the following: When on varying a 
parameter the stochastic instability sets in, the continued use of an algo- 
rithm describing the evolution of the system prior to the onset of chaotic 
behavior may no longer be appropriate. In this event it may be more prac- 
tical to take advantage of the stochastic nature of the system and use more 
or less conventional statistical methods. In order to effect such a change in 
computational methods it is necessary to detect the change from orderly 
behavior of the system to chaotic behavior. Theoretically, the sensitivity 
of the system to initial conditions accompanies the occurrence of positive 
Lyapounov exponents, i.e., numerical indices of the asymptotic rate of 
divergence of initially nearby system trajectories. Since the Lyapounov 
exponents are defined as time asymptotic quantities, straightforward com- 
putation of these quantities requires a considerable computational effort, 
including a calculation of the evolution of two initially nearby trajectories, 
or the simultaneous integration of the associated variational equations. 

An important tool for studying the properties of dynamical systems 

57 



in the chaotic regime is the first return map, sometimes referred to as 
the Poincar6 map. Such a map displays in a graphical manner various im- 
portant aspects of the attractor. Unfortunately, the simplicity of generat- 
ing such a map is restricted to systems no larger than three dimensional, 
because of the self-evident difficulty of making multidimensional graphic 
displays. As the study of higher-dimensional dynamical systems advances 
there -will be an ever more urgent need for a higher-dimensional equivalent 
of the first return map. 

As mentioned earlier, in the chaotic regime it may be more appropriate 
to describe the properties of the system trajectory in statistical terms than 
in terms of a trajectory evolution. In order to obtain such a statistical 
description it is necessary to have appropriate information about the in- 
variant measure associated with the given dynamical system. With few 
exceptions such measures cannot be derived a priori but must be obtained 
from detailed calculations of the trajectories. A finite machine computa- 
tion of the system trajectory will inevitably introduce some errors that 
may be all the more serious in the chaotic regime because of the previously 
mentioned exponential divergence of nearby trajectories. There is then the 
question of how accurate the computation of a chaotic trajectory must 
be in order to yield enough information for constructing the appropriate 
invariant measure. 

While the above examples have been drawn from dynamical sys- 
tems representable by ordinary differential equations there are other sys- 
tems, e.g., those described by partial differential equations and by integral 
equations, whose study is likely to be replete with similar computational 
difficulties. 



3.5.3 Symmetry Breaking 

Bifurcations commonly arise in connection with a loss, or breaking of 
symmetry. In such cases the extra structure of the problem symmetry 
may simplify the analysis. We give an example from fluid mechanics, A 
classic experiment concerns flow between two rotating vertical cylinders. 
For small angular velocities of the cylinders the flow is laminar and is called 
Couette flow. As the angular velocity is increased the vertical translation 
symmetry is broken and axisymmetric Taylor cells appear. These cells 
resemble a row of adjacent smoke rings, each one rotating in a sense 
opposite to its immediate neighbor. A further increase in velocity breaks 
the axial symmetry, and the Taylor cells become wavy with a time periodic 
shape. Eventually the flow becomes turbulent. Specific experiments on 

58 



these systems are a classic topic in fluid mechanics. However as often 
happens in science, one cannot see important phenomena in the absence of 
a carefully thought out theory. Recently a number of precise experiments 
have been conducted concerning the role of strange attractors. Routes to 
turbulence and, hi particular, wavy Taylor cells were carefully observed. 
The results confirmed some of the theoretical predictions but not all of 
them. One theoretical proposal was that generic limiting sets, known as 
strange attractors, would dominate turbulent flow patterns. This proposal 
was supported by the fact that strange attractors were observed in low- 
mode approximations to fluid flows. More refined calculations with more 
modes included indicate an absence of these strange attractors, but the 
phenomenon is still indicative of the highly complex solution manifolds 
that can arise hi nonlinear problems. 



3.6 ILL-POSED INVERSE PROBLEMS 

The notion of a well-posed problem is due to Hadamard: a solution must 
exist, be unique, and depend continuously on the data. The term "data" 
can have a variety of meanings; in a differential equation it could include 
any or all of the following: boundary values, initial values, forcing term, 
and even the coefficients in the differential equation. Since data cannot be 
known or measured with arbitrary precision, it was felt for a long time that 
real physical phenomena had to be modeled by mathematically well-posed 
problems. This attitude has changed considerably in recent years, and it 
is now recognized that many applied problems are ill-posed, particularly 
when they require numerical answers in the presence of contamination of 
the data. 

El-posed problems often arise in the inversion of well-posed problems. 
Consider, for instance, a well-posed problem that smooths the data or 
attenuates its high frequencies. The inverse problem, in which the role of 
data and solution are interchanged, will then be ill-posed. A simple but 
important example is the Fredholm integral equation of the first kind 



/ k(x, y)u(y)dy = v(x), < x < 1 
Jo 

or, in operator form 

Ku = v 

Assuming the kernel to be continuous on the closed unit square, the 

59 



Riemann-Lebesgue lemma gives 

yi 
lim / k(x t 2/)sin nydy = 

n ^oo JO 

so that K attenuates high frequencies and thus transforms widely different 
functions u into approximately the same v. If v is regarded as the data, the 
solution u of the integral equation, therefore, cannot depend continuously 
on the data. Indeed, it is not necessarily uniquely solvable, either take, 
for example, 

N 



If, unaware of these difficulties, one attempted to solve the integral equa- 
tion by discretization, one would find that the corresponding matrix prob- 
lem is singular or ill-conditioned (and the finer the discretization, the more 
ill-conditioned the matrix problem). 

More general inversion problems can also be reduced to an equation 
of the form Ku = v, where K is a continuous transformation, v is the 
data, and u is the solution being sought. It may happen that u itself is 
not the quantity of principal interest but rather some functionals of u such 
as some of its moments or its values at a few specified points. 

Thus, there are three pieces of information that are central to the 
numerical resolution of an inverse problem: (1) the model M, representing 
the equation involving a mapping K between appropriate spaces; (2) the 
observation operator 0, representing the measurements that can be made; 
for instance, we might have O(w) = {v(xi)-,i = l,2,.,.,n}; (3) the intel- 
ligence operator J, which specifies the information we wish to extract from 
the solution. For instance, we might have J(u) = {fx k u(x)dx; k = 0, 1, 2}. 

Most of the existing approaches focus almost exclusively on the model 
M, taking account of a priori information about the solution such as 
smoothness, positivity, and bandwidth. The Tychonov regularization 
method is of this kind. By restricting K to a suitable subspace it may 
be possible for the restriction of AT to be one to one with a continuous in- 
verse. It might be more appropriate to study the triplet M,0, J for different 
choices of O and J. One goal of such a new approach would be to classify 
and quantify ill-posedness by developing comparison principles and order 
structure. 

Hi-posed problems with their attendant numerical difficulties abound 
in practice. For instance, in scattering from obstacles one may wish to 
determine the shape of an object (or its surface impedance, if its shape is 



60 



known) from far-field measurements. Coefficient identification problems 
arise in many contexts, one such problem being the determination of the 
sound speed in subsurface media from measurements of the field at or 
near the surface. Another class of ill-posed problems is associated with 
image reconstruction (tomography), the medical applications of which are 
now widely used (CAT scanners, for example) (see Section 2.12). More 
recently, similar ideas have been applied to the nondestructive testing of 
mechanical structures to detect cracks in fuel rods, weaknesses in rotor 
blades of jet engines, and faults in screws in ships. Optimum filtering 
and inverse problems in Fourier optics (restoring data designed by a band- 
limited filter, for example) are other areas of current research. 



3.7 EFFECTIVE MEDIA 

The study of "bulk" or "effective" parameters for composite media is of 
fundamental importance. Depending on the particular application area it 
may be relevant to consider a periodic (or an almost periodic) or a random 
formulation. 



3.7.1 Homogenization, the Deterministic Approach 

Many problems of physical interest involve several length scales. As an 
important example we mention the study of composite materials in struc- 
tural mechanics. Owing to a particular manufacturing process, a distinct 
structure, e.g., periodically or almost periodicity, is often present. 

Homogenization is an approach for deriving the macroscopic 
properties of the material from the known microscopic ones. A variation 
of this is to replace a complicated geometrical configuration by a simpler 
one, e.g., to replace a framework by a plate or to smooth out a rough 
surface. 

Homogenization may be applied to linear and nonlinear problems and 
can provide qualitative information about physical macrolaws as well as 
good approximations to the various parameters appearing in these laws. 

The most important mathematical tool for the deterministic approach 
is some form of asymptotic expansion. The theoretical results concern the 
limiting behavior when certain of the relevant scales become very small. 



61 



One is likely to obtain different homogenized formulations from different 
limiting relations between the length scales. 

The numerical solution of a problem introduces a new length scale: 
the mesh size in a finite-element method, the step size in a finite-difference 
method, or the wavelength in a spectral method. Different limiting rela- 
tions between this new length scale and the original physical length scales 
usually lead to different results. In this sense the numerical treatment has 
to be considered simultaneously with the homogenization process. One has 
to design algorithms that will adaptively select the correct homogenized 
formulation and discretize it appropriately. 

Whereas much theoretical work has been done within the last ten 
years analyzing the effects of multiple scales in continuous media, the 
study of numerical discretizations of such problems is still very much in 
its infancy. 



3.7.2 Analysis of Random Media 

Assessing the effect of random fluctuations in the coefficients of a par- 
tial differential equation is a basic mathematical problem that arises con- 
stantly in science, applied science, and engineering. Finding effective con- 
ductivities of composite conducing materials such as soil or a metal alloy, 
finding effective fluid equations for flow in porous media (Darcy's law), 
determining the rate of sedimentation of particles in a fluid, and many 
other problems are problems that may require analysis of random media. 
In addition to the quantitative aspects of the problem, many interesting 
qualitative questions can be posed as well. For example, what is the nature 
of the spectrum of the Schrodinger equation with a random potential? It 
has been shown that in one dimension, for a large class of random poten- 
tials, the spectrum is always discrete. For randomness of large magnitude, 
it was recently shown that there is no diffusion, without restriction on the 
dimension of the space. 

From the applications' viewpoint, one can frequently model ade- 
quately a random medium, such as a suspension, for example by a con- 
tinuum with suitable constitutive properties. The continuum equations 
may be linear or nonlinear, and the constitutive laws may be known 
only qualitatively from experimental data. In such a context, theoreti- 
cal investigations are useful and necessary in order to understand how 
the phenomenological continuum equations arise from the known micro- 
scopic Jfrueta* One can then find mathematical characterizations for 
the relevant constitutive laws that can lead to interesting conclusions. 

62 



Examples of this arise frequently as, for instance, the determination of 
bounds for the effective conductivity of composites. 

The effective media ideas of Maxwell in the last century dominated 
theoretical calculations for a long time. In the last two decades, press- 
ing technological needs have caused a major expansion in materials re- 
search. In attempting to achieve a mathematical understanding of the 
methodological basis for effective media calculations that abound, one finds 
a lack of theoretical foundations. 

Mathematical methods in random media are drawn from analysis, 
probability theory, asymptotic methods, and differential equations. The 
goal is to develop tools that bridge the gap between microscopic and 
macroscopic descriptions, give qualitative information about constitutive 
laws, and determine when residual stochastic effects remain and how they 
can be characterized. In addition, it is of interest to find important specific 
problems on which more detailed analysis, including numerical analysis, 
can be carried out so that these problems can serve as benchmarks in the 
theoretical development. 



3.8 VALIDATION, ERROR ASSESSMENTS, AND 
SENSITIVITY ANALYSIS 

The results look good, but are they really right? The question is often 
not answered satisfactorily. Indeed, often too little attention is paid to 
the difficult topic of model and code validation. Once the results look 
plausible, we are often either unable or unprepared to take the valida- 
tion process further. The validity of a model depends both on having a 
proper physical model (Do the governing equations adaquately represent 
the physics?) and on having an accurate computational representation of 
the governing equations. The mathematician has a responsibility in both 
areas. He should help determine the "well-posedness" of the models and 
from a mathematical point of view help the physical scientist determine 
the appropriateness of the model. Given that the equations are proper, the 
computational mathematician must be sure that the numerical procedures 
used accurately approximate the solutions of the governing equations. 

Frequently modeling is done in conjunction with experiments, and 
those results are used to validate the model. Comparison with experiments 
simultaneously tests both the validity of the governing equations in the 
model and of their numerical solution. In this case the source of any 
discrepancy is not easily isolated. A reasonable goal should be to validate 
the numerical procedures and the physical models separately so that a 

63 



model can be developed more confidently. The design of reliable numerical 
error estimates for the computational methods is an important step. 



3.8.1 A Posteriori Error Estimates 

An important current research topic is the development and application 
of a posteriori error estimates. Such procedures are valuable for models in 
which differential equations are approximated by their discrete analogs on 
a mesh network. When a posteriori estimates of the error associated with 
the discretization of the continuous model are possible and they often 
are then potentially it is possible to control the errors. Adaptively moving 
the mesh network to control or mimimize the error is one application of 
this approach. With a posteriori error estimates, it should be possible, 
in principle, to give a strict bound for the error on completion of the 
computation. 

The selection of the specific accuracy requirements depends strongly 
on the goal of the computation. Often it is desired to obtain detailed 
information about the solution itself; in other cases, the main focus is the 
value of a specified functional of the solution, as, for example, a stress in- 
tensity factor in fracture mechanics or a drag coefficient in fluid dynamics. 
Other goals may be the determination of certain critical data, such as 
collapse points or buckling points and their associated loads in structural 
mechanics. Sometimes special techniques, e.g., variational techniques, can 
yield highly accurate estimates of particular functionals. 

In connection with most of these goals we are interested in quantita- 
tive results that have a desired accuracy. For this the error has to be 
defined, in that a family of exact results has to be specified with which 
the computed data are to be compared, a norm has to be prescribed in 
terms of which the error is to be measured, and some procedure has to be 
established for estimating the error. 

Such error estimation capabilities are certainly important hi many 
applications, provided they can be guaranteed to be reliable. For example, 
for the many types of certification computations required hi the design of 
complex structures or nuclear plants, the availability of reliable estimates 
of the accuracy of the computed data is obviously important. In other 
cases, such estimates may reduce the total design effort and avoid un- 
necessary overdesign. At the same time, the availability of effective er- 
ror estimates introduces the possibility of applying adaptive techniques to 
structure the computation to achieve a desired error tolerance at minimal 
cost or to provide the best possible solution within an allowable cost range. 

Many of the theoretical studies of solution algorithms for classes of 

64 



differential equations provide for some a priori error estimates, usually 
of an asymptotic type. For the practical error assessment these a priori 
bounds are rarely computable or very accurate. Thus, one is led to the 
necessity of developing a posteriori error estimates that utilize information 
gained during the computation itself. 

For the finite-element solution of certain classes of elliptic boundary- 
value problems, some computable and reliable a posteriori error estimates 
have been developed and analyzed in recent years. Most of these a pos- 
teriori estimates are based on a local analysis. For a given mesh A consist- 
ing of elements AI, . . ., A n and with a corresponding approximate solution 
on that mesh, an error indicator rjj is associated with the yth element A J( 
These r\j have to be computable in terms of information about the prob- 
lem and the approximate solution on Aj and, at most, on the immediate 
neighbors of that element. On the basis of the indicators r}i,...,r) n an 
error estimate e(A) is then constructed. Of course, the T)J as well as e(A) 
depend on the chosen norm. The theoretically important question is then 
the relationship between e(A) and the norm of the actual error e(A). The 
effectivity of the estimation may be judged by the effectivity index 9 = 
(A)/||e(A)||. In practice, it is usually more important for 6 to be close 
to 1 than that < 1. Moreover, it is essential that 6 converges to 1 when 
||e(A)|| tends to zero, so that for an accuracy of, say, 5 to 10 percent the 
value of \0 1| is expected to be less than, say, 0.1 or 0.2. 

The requirement of designing error estimators with these properties 
for realistic classes of problems and various different norms certainly repre- 
sents a demanding research task. The results available so far suggest that 
such error estimators can be developed at least for linear problems. For 
nonlinear problems the situation is in an embryonic stage although some 
results available for model problems in one space dimension indicate that 
estimators for the error along continuation paths and for the location of 
critical points are computationally feasible. There is certainly considerable 
need for concentrated research in the general area, 



3.8.2 Other Validation Measures 

Checking for mesh dependencies and convergence rates ie another way to 
help validate a model. If a method is supposed to be, say, "second order," 
then as the mesh intervals are halved, the errors should be reduced by a 
factor of 4. If this does not happen according to one's error estimate, then 
the method does not have the desired order yet and may be in error. Also 
one should continue to refine the mesh until the results are independent of 
refinement to within the desired accuracy. Solutions can also be checked 

65 



for sensitivity to mesh orientation. The solution should not be aligning 
itself with some special property of the mesh network. 

Numerical methods can also be validated by comparing them with 
analytic solutions. Often this is possible by choosing limiting cases where 
an asymptotic analysis applies. In some cases complex constitutive laws 
can be replaced with constants so an exact solution is possible. Higher- 
dimensional models can sometimes be validated by comparison with pre- 
viously confirmed lower-dimensional results. Bounds on solutions are 
another way to help verify a model. Often conservation laws or minimum 
and maximum principles apply, or the solution is known to approach a 
previously known equilibrium or steady-state condition. These properties 
can be used as a basis for validation. 

To the extent possible one should always write computer codes in 
a modular fashion such that each part of the model can be validated 
separately. Even though this is just good structured programming prac- 
tice, and not necessarily related to mathematics, it is an important aspect 
of model validation. 



3.8.3 Sensitivity Analysis 

In addition to the independent and dependent variables, most models also 
depend on certain physical parameters. Unfortunately, those parameters 
are often not known accurately. Sensitivity analysis is a systematic 
means to quantify the relationship of model parameters and model results. 
Doing the sensitivity analysis requires solution of an additional system 
of differential equations. These equations are a formal statement of 
the relationship between the dependent variables of the system and the 
parameters. The results are given in terms of a matrix of partial deriva- 
tives of the dependent variables with respect to the system parameters. 
One way to think of the analysis is as a method to provide theoretical 
"error bars" for the model. 

New methods are being developed to solve these sensitivity equations 
quite efficiently. The methods rely on the observation that the sensitivity 
equations are linear, regardless of the nonlinearities in the original model. 
The method solves differential equations for the Green's function of the 
sensitivity equations, and then the sensitivity coefficient matrices are com- 
puted by quadrature for the various imhomogenous terms corresponding 
to the parameters. These methods have been successfully applied to prob- 
lems in chemical kinetics. 



66 



4. NUMERICAL METHODS 



4.1 DISCRETIZATION METHODS 

Many important physical problems are modeled by boundary or initial- 
boundary-value problems. In this modeling, the physical state under 
consideration is characterized by a function u, which is the unique solution 
to the boundary or initial-boundary-value problem. Thus a major part of 
the analysis of the physical problem is the determination of u or some of 
its derivatives or some functional of it. 

In nearly all important problems the determination of u is an infinite- 
dimensional problem in the sense that u does not lie in an explicitly 
known finite-dimensional space or, alternatively, cannot be expressed in 
terms of a set of explicitly known functions by means of a finite set of 
parameters. Thus u, cannot be "completely" represented on a computer, 
and it is necessary to resort to some type of approximation or discretization 
method. In essentially all discretization methods one attempts to construct 
a function, u &pprox , which on the one hand is "close" to u and on the other 
is characterized by a finite set of parameters, and so can be calculated. 

The aim of finding ii approx that is "close" to u is made precise by 
requiring that \\u - u &pprox \\ < r (or that \\u - u appr ox || < r||u||), where 
1 1|| is a physically relevant norm and r is a physically relevant tolerance. 
The goal is thus to find u app rox satisfying this error criterion with the 
least expenditure of computational effort. The selection of a discretization 
procedure is influenced by a number of considerations, the most important 
of which we now list: 

1. The goals of the analysis of the physical problem. 

2. Known mathematical properties of the physical problem and the 
algorithm. 

3. Hardware considerations, e.g., the availability of parallel process- 
ing. 

4. Computer-science considerations, e.g., data-management require- 
ments, 

5. Restrictions on computation time and expense. 

67 



As indicated above, the exact solution is approximated by a function 
u a pprox , which is expressible in terms of a set of explicitly known (basis) 
functions by means of a finite set of parameters, which are determined 
in the computation. One broad classification of methods is in terms 
of the nature of the basis functions, namely (a) those involving global 
basis functions having global support and (b) those involving local basis 
functions having small support. Another classification is based on tne 
extent to which the method is adaptive. Adaptive methods refine and 
modify themselves on the basis of partially completed computations. 

Of course, these classifications are not precise, and there are methods 
possessing the different properties in various degrees. We now turn to a 
more detailed discussion of the most commonly used discretization proce- 
dures. 



4.1.1 Finite Differences 

One of the most frequently used discretization methods is the method 
of finite differences. The central idea is to replace each partial deriva- 
tive occurring in the differential equation in the underlying boundary or 
initial-boundary-value problem by an approximating difference quotient, 
For example, the first-order derivative du(x,y)/dx may be replaced by 
the forward difference [u(x + h,y) -u(x,y)]/h or the backward difference 
[u(x,y] - u(x - fr,y)]//i, whereas the Laplacian Au = u xx + Uyy may be 
replaced by the five-point difference operator 



u(x + h.y} + u(x. v + h) + u(x- h. 



This replacement of all derivatives in the differential equation by ap- 
propriate difference quotients leads to a system of equations, called 
difference equations, for the numbers Uij, which are to approximate the 
values of the exact solution u(x,y] at the finite-difference mesh points 
(yhi kh), j, k = 0, 1, +2, . . ., where h is a small positive number, called the 
mesh parameter. Nonuniform meshes may be considered as well, but at 
considerable complication in the form of the difference quotients in the 
resulting difference equations. 

There are a number of important questions that arise in connection 
with the study of difference methods, namely questions of 

68 



1. Accuracy of the method; 

2. Stability of the method; 

3. Efficient solvability of the finite-difference equations; 

4. Robustness of the method with respect to the input data (e.g., 
coefficients, forcing functions, meshes); and 

5. For nonlinear problems, the separation of legitimate approximate 
solutions, which correspond to an exact solution, from spurious ap- 
proximate solutions, which do not correspond to any exact solution. 



Finite-difference methods are an example of the local basis class of 
approximation methods. They can be adaptive. Adaptivity is usually 
introduced via adaptive mesh selection: the mesh chosen at any stage in 
the computation is based on the previous computations. 

Much important work on difference methods remains to be done. We 
mention in particular the construction of effective difference methods for 
nonlinear problems, in particular for those problems whose solutions have 
shocks. 



4.1.2 Variational Methods of Discretization 

We will now discuss a class of discretization methods based on variational 
formulations of the physical problem under consideration. As noted above, 
the problem of calculating the exact solution u is infinite dimensional in 
the sense that u is only known a priori to lie in an infinite dimensional 
space, say H. For a linear problem, a variational method of discretization 
consists of (a) a finite-dimensional space S cH, called the trial space, 
in which the approximate solution is sought, (b) a finite dimensional test 
space V, and (c) a bilinear form B(u, v) defined on H X V. The approximate 
solution is then determined by requiring that 



u 



approx 



-B(ii a pprox > i>) = B(u, 11), for all v 6 V 

where for veV , B(u, v) is computable from the data of the problem without 
knowing u, Since and V are finite dimensional, u app rox can De calculated 
by means of the solution of a system of linear equations if the original 
equation is linear. Usually, for nonlinear problems, B is nonlinear in u, 
and the resulting system of equations becomes nonlinear. 

69 



The approximate solution depends on the selection of S,V, and B. 
For any problem there exists a wide variety of variational methods of 
discretization, i.e., a wide variety of choices for S, V, and B. The rational 
selection of S, V, and B is a central problem. 

Variational methods can be of either the local-basis or global-basis 
type. Of those of global-basis type we mention the various versions of the 
spectral method and the p version of the finite-element method. Finite- 
element methods, considered in more detail below, and collocation methods 
are typically of the local-basis type. We also note that variational methods 
can be adaptive. 



4.1.3 Finite-Element Methods 

Finite-element methods arise if for 3 and V we choose spaces S h and V h 
of piecewise polynomials of fixed degree. That is to say, the underlying 
spatiaMomain for the problem is broken up into small geometric pieces, 
called "elements," whose size is measured by a parameter h. The functions 
in both S h and V h are then restricted to be polynomials in x and y on 
each piece but allowed to be different polynomials on the different pieces. 
In designing finite-element methods, i.e., in selecting S ht V h , and B, one 
attempts to achieve approzimability, stability, and systems of equations 
that can be solved effectively. 

Approximability here refers to the ability of the space S h to ap- 
proximate the unknown solution u. The solution u is unknown a priori, 
and often only the information u e H is available. In such situations 
6 h has to be selected so that every function in H can be approximated 
sufficiently well by one in S h . However, an approximation based on a few 
large elements can provide additional information on u, which can be used 
in turn to refine those elements. This type of adaptive element selection 
is especiaUy important for problems with sharply varying solutions (see 
section 4.2). 

In the choice of the bilinear form B one has, in effect, the freedom to 
choose a variety of variational principles, many of which have a natural 
physical connection with the original problem. For instance, the stan- 
dard Rita method is based on the principle of minimum potential energy. 
An alternate variational principle is the complementary energy principle. 
Approbation methods based on this principle are referred to as com- 
plementary energy or equilibribum methods. These methods involve a 
constraint that is usually difficult to satisfy. One way to circumvent this 
difficulty is to treat the constraint by means of a Lagrange multiplier. This 



70 



leads to the so-called mixed methods. They appear to be promising for 
many important problems and have recently received a large amount of 
attention. 



4.1.4 Transformation Methods 

In the transformation or pseudospectral method, the discrete approxima- 
tion u is first mapped by a transformation of the form 



into the m-dimensional function space of the coefficients, a,j. The basis 
functions fa and the transformation are chosen so that T and T~ l are 
fast (order of m log m operations) and so that differentiation D is simple 
in transform space. The derivative approximation can then be written as 



dx 

Some common transforms are based on the fast Fourier transform, where 
the fa are trigonometric functions or Chebyshev or Legendre polynomials. 
Selecting the <t> 3 - as piecewise polynomials with compact support, such as 
the B-splines, is another good choice. By choosing the transformation 
to incorporate some crucial property such as the periodicity or symmetry 
of u one can improve the accuracy of the method for a fixed number of 
basis functions. This can sometimes best be done by choosing the basis 
functions close to the eigenfunctions of the differential equation. 



4.1.5 Semidiscrete Methods 

When solving a partial differential equation one sometimes discretizes with 
respect to some but not all of the variables. For example, for the diffusion 
equation governing the cooling of a hot rod, one may discretize with respect 
to the space but not the tune variable, thereby replacing the original 
partial differential equation by a system of ordinary differential equations. 
Such an approximation method is referred to as a semidiscrete method or 
as the method of lines. Semidiscrete methods may be used for hyperbolic 
as well as parabolic equations. 

71 



One has, of course, eventually to discretize with respect to the time 
variable as well, Semidiscrete methods are based essentially on the view 
that very effective time discretization methods are available (in the form of 
preprogrammed software packages for the solution of ordinary differential 
equations) and that the spatial discretization is the main concern. An 
alternative point of view is to consider both discretizations simultaneously. 
Such fully discrete methods have been analyzed and tested extensively. 

Among important research problems for semidiscrete methods we 
mention the problem of adaptive mesh selection for the spatial discretiza- 
tion. 



4.1.6 Method of Characteristics 

This is a method for hyperbolic equations, particularly for those involving 
only one space dimension. In these equations the solution at some point 
in space-time depends primarily on its values along a fixed, finite number 
of curves (characteristics) going back in time from the given point. The 
approximations to these values are determined from difference equations 
that are closely related to the characteristics of the differential equations. 
This method has a natural generalization to quasi-linear second-order 
equations in two independent variables. It is especially important for 
problems whose solutions have shocks. 



4.2 ADAPTIVE GRID METHODS 

For realistic problems it is rarely feasible to design numerical processes that 
are rehable and accurate at a reasonable cost and yet that do not utilize 
some form of adaptivity. Put simply, most two-dimensional and virtually 
aJI three-dimensional problems are undercomputed without this technique. 
This statement will almost certainly remain true after the next generation 
of computers is available. The adaptive approach is to distribute the com- 
putational effort nonuniformly, so as to concentrate on the most singular 
parts of the solution or the most important parts of the solution from the 
point of view of critical design parameters. Correspondingly one must 
give less computational effort to the regular or less critical parts of the 

SL T 'I** SeCti ^ 4 ' 9) - M ** 8ame time > >-P**VB approaches 
may also lead to a simplification of the input data needed for the program 



72 



and hence free the user of part of the drudgery typical in preparing such 
input and in having to make the many a priori decisions required by most 
of today's programs. 

The goal of adaptive grid methods is to choose a grid that is par- 
ticularly refined, or that is aligned or oriented optimally, with respect to 
the solution in regions of space and time that are critical for solution ac- 
curacy. Thus adaptive grids utilize local mesh refinement or optimal local 
mesh orientation. 

The simplest adaptive grid is a preliminary, analytically determined 
coordinate transformation. For example, in the transformed coordinates 
the problem may be approximately independent of one variable or other- 
wise simplified. The next strategy is to choose a numerically deter- 
mined coordinate transformation. Typically, in two dimensions the coor- 
dinate transformation is obtained by the solution of a subsidiary partial 
differential equation. The resulting grids may be expected to give both 
improved mesh refinement and mesh orientation. The method is some- 
what problem dependent and can give rise to discretization errors in the 
mapping of solution variables between the various grids. 

Lagrangian grids for tune-dependent problems are adaptive for 
material interface singularities, because the grid points move with the 
material particles. Since these well-established methods also develop 
rezoning and mesh entanglement problems, they provide a reservoir of ex- 
perience concerning the difficulties associated with other evolving, adaptive 
algorithms. 

A refined grid can be constructed without recourse to a coordinate 
transformation. In response to a critical solution parameter or solution 
error criteria, selected regions of space can be flagged, preferred orientar 
tions selected, and refined subgrids introduced locally. Then, small time 
steps are chosen on the finer subgrid, and an interpolation problem must 
be solved to blend fine and coarse grid solution values. Finally, the con- 
struction is recursive, so that three, four, etc. levels of refinement can be 
introduced automatically, in response to some a posteriori error estimate 
on the next coarser grid. Precisely defined, reliable error estimators appear 
to be central to the design of effective adaptive procedures. 

In the context of finite elements for elliptic problems there is a devel- 
oped theory for a posteriori error estimates (and, hence, for adaptivity), 
which is based on local analysis. For a given mesh made up of elements 
AI, . . ., A n and with the aid of the corresponding approximate solution, an 
error indicator r)j is computed for each element Aj. For certain classes of 
problems it has been proven that asymptotically the errors become optimal 
when all the r\j become equal. This equilibration principle provides the 
basis for the control of the adaptive process. In essence only those elements 

73 



A.,- are refined for which the error indicators are too large in comparison 
with those of the other elements. And elements that are unnecessarily 
small are combined with their neighbors. The study of suitable algorithms 
for this has only recently begun. For example, some results have been ob- 
tained for algorithms based on the assumption that during the refinement 
process none of the elements could be combined without increasing the 
maximum value of the error indicators. 

In tune-dependent problems, differential equations can be introduced 
for the ever-changing optimal location of the grid nodes. These equations 
are then solved simultaneously with the original partial differential equa- 
tion, leading to what are known as moving mesh methods. 

Alignment of the grid without refinement is also possible, if fixed 
or moving boundaries or interior interfaces are specified as part of the 
problem. By alignment (without refinement), a regular grid index structure 
can be preserved. Maintaining the grid structure has the advantage that it 
potentially allows fast iterative methods, such as fast Fourier transforms, 
to be used as part of the solution algorithm. 

Adaptive grid techniques have been applied successfully to a range of 
time-independent problems. The newer time-dependent methods need to 
be developed to the point where they can be applied to meaningful two- 
and especially three-dimensional problems and compared with alternative 
methods. An important question requiring further attention is the con- 
struction of efficient, computable a posteriori error estimates for realistic 
classes of problems. Specifically, even for steady-state problems, there is 
the question of the validity of the equilibration principle mentioned above 
as well as the design and analysis of suitable adaptive control laws to 
implement this principle. This latter problem may require examination 
and incorporation of results in such fields as automatic control theory, 
artificial intelligence, and learning processes. It also raises the problem 
of the choice of appropriate refinement techniques that produce meshes 
with desirable properties. Last but not least there is the question of data 
management, which must be handled successfully to control the vastly in- 
creased algorithmic complexity associated with adaptivity and to achieve 
computational efficiency. 



74 



4.3 METHODS FOR SOLVING DISCRETIZED DIFFERENTIAL 
EQUATIONS 

Finite-difference and finite-element discretizations of partial differential 
equations usually give rise to large systems of equations in which each 
unknown is coupled to only a few of the other unknowns. Systems with 
tens or hundreds of thousands of equations are relatively common. For 
time-dependent problems these systems arise from the use of implicit 
time discretizations. For sufficiently fine grids, the numerical solution of 
these systems consumes a major part of the computer time for an entire 
simulation. 

Most nonlinear systems are solved by some form of iterative method 
based on linearization, such as Newton's method. At each step, these 
methods result in large sparse linear systems. In many cases, iterative 
methods converge only if the initial guess is sufficiently close to the solu- 
tion. 

There are two basic approaches to solving large sparse linear systems 
of equations: direct sparse matrix methods (i.e., some form of Gaussian 
elimination that takes advantage of the location of most of the zeros) 
and iterative methods, where the sparsity makes each iteration relatively 
inexpensive. No method is best for all problems. For many problems 
a combination of methods is attractive. Usually this takes the form of a 
block iterative scheme using a direct method to solve the subsystems whose 
diagonal blocks are matrices. Combination methods are of particular 
importance for the solution of linear systems arising from coupled systems 
of partial differential equations. 

Direct methods are relatively well understood today, and a number 
of excellent implementations of them exist for serial computers. Direct 
methods vary in the extent to which they take zeros into account. The 
simplest, nontrivial approach is band elimination, which takes account 
only of those zeros occurring outside the band determined by the extreme 
nonzeros of the matrix. The most complex approach is general sparse 
elimination in which all the entries that remain zero during the alimrnflr 
tion are taken into account. For systems that do not require pivoting 
(e.g., those having symmetric, positive definite matrices) there is a sym- 
bolic preprocessing phase in which the location of these zero entries is 
calculated. There exist good techniques (the nested dissection ordering 
and the minimum degree ordering) for arranging the unknowns and the 
equations so as to minimize the zero fill-in during elimination. For model 
problems it has been shown that the nested dissection ordering provides 
asymptotically optimal results with respect to work and storage. The 

75 



minimum degree algorithm is a valuable heuristic approach that is com- 
petitive in practice but that has defied analysis. For systems that require 
pivoting for numerical stability one cannot compute the zero structure a 
priori. Moreover, the ordering of the unknowns and equations to minimize 
zero fill-in mil usually be significantly altered. 

Some of the strong points of direct methods are as follows: (1) They 
provide an exact answer (modulo round-off error) to the linear system with 
a fixed number of operations independent of the system's condition num- 
ber. Most production structural analysis packages such as NASTRAN use 
some form of direct method, even for three-dimensional problems. In these 
applications (many of which are based on fourth-order elliptic problems) 
it is necessary to use higher-order precision because of conditioning prob- 
lems. (2) For problems with two dependent spatial variables, their execu- 
tion time is often shorter than the execution tune for iterative methods, 
especially for problems of moderate size. Some of their principal disad- 
vantages are the following: (1) they require considerably more storage than 
iterative methods, even for two dimensions; (2) they will almost always be 
noncompetitive with iterative methods for three-dimensional problems in 
terms of running time; and (3) except for the simpler methods such as 
band elimination (and to a lesser extent profile or envelope elimination) 
they do not vectorize well. In fact, owing to the necessary indirect ad- 
dressing involved, there are as yet no efficient implementations of general 
sparse direct matrix methods for the CRAY-1 or CDC CYBER-205 super- 
computers. 

Except for structural problems, iterative methods are commonly used. 
Classical iterative methods, such as successive overtaxation, Chebyshev 
semi-iterative methods, and such newer methods as the preconditioned 
conjugate gradient method, are fairly well understood for symmetric, posi- 
tive definite systems, and they are easy to implement. However, the situa- 
tion is not so bright for nonsymmetric or indefinite systems, though in 
practice classical iterative methods may converge rapidly with a clever 
choice of the iteration parameters. Nonsymmetric systems with indefinite 
symmetric parts are especially difficult to solve, and no satisfactory general 
algorithms are known at this time. Such problems arise in the simulation 
of secondary and tertiary thermal recovery techniques for petroleum reser- 
voirs (see Section 2.6). 

Some of the principal advantages of iterative methods are the follow- 
ing: 



1. They tend to require minimal storage (proportional to the number 
of unknowns); 



76 



2. They are reasonably fast for a wide range of problems. Moreover, 
the number of iterations required is independent of the number of space 
dimensions in the underlying partial differential equation (but not of their 
order, or of the mesh size); 

3. They can take advantage of good initial guesses to the solution, as 
would be available in time-dependent or nonlinear problems; and 

4. Many of them vectorize reasonably well on supercomputers. 



Some of their disadvantages follow: 

1. Mathematically rigorous stopping criteria may be difficult to for- 
mulate, e.g., for linear systems with matrices that are not symmetric, 
positive definite; 

2. Many of the methods require a selection of iteration parameters, 
and the performance of the methods depends critically on such a selection 
(this difficulty is being overcome somewhat by the relatively new adaptive 
methods and the preconditioned conjugate gradient-type methods); 

3. The interaction between linearization and iteration is not well 
understood, especially for discretizations of coupled systems of partial 
differential equations; and 

4. Nonsymmetric or indefinite problems may cause great difficulties 
for iterative methods. 



The relatively new multigrid iterative method combines the well- 
understood behavior of a given iterative technique with the fact that an 
underlying differential equation is being solved approximately. It alter- 
nates iterating, toward the differential equation's solution, on fine and 
coarser subgrids of the spatial network, with the goal of performing no 
more computational work on the finer (hence expensive) grids than is ab- 
solutely necessary. In many cases of practical interest, such as in neutron 
diffusion in complex environments, the technique yields sufficiently ac- 
curate solutions to the equations in a computational time proportional to 
the number of unknowns. This has been a long-sought-for goal in the 
approximation of elliptic equations in more than one space dimension. As 
an iterative method it also has a natural extension to nonlinear equations; 
and its logical structure, together with the already necessary calculation 
of residual errors, points toward the incorporation of adaptivity for the 
nesting spatial grids. 



77 



4.4 CONTINUATION AND HOMOTOPY METHODS 

For more than a century the so-called continuation principle has proved 
to be an important tool in mathematical theory. For example, early uses 
date back at least 100 years to H. A. Schwarz and his work on con-formal 
mappings; then, in the early part of this century, J. Hadamard and M 
Levy applied these techniques in connection with the inversion of nonlinear 
mappings, and it is also a basic tool in J. Leray and J. Schauder's work on 
differential equations. But it was essentially only in the early 1950s, with 
the advent of automatic computing, that continuation approaches began 
to be used in numerical applications. Since then they have grown into an 
extremely powerful technique for the numerical solution of wide classes of 
nonlinear problems. 

One of the problems to which continuation techniques are applied 
concerns the solution of a nonlinear equation F(x) = in some space 
X. In order to compute a specific solution x* X a possible approach is 
to imbed the equation into a family of equations H (x, t) = 0, < t < I, 
for which there exists a computable solution path x = x(t), 0<t<l, in 
X beginning at a known point z(0) and ending at the desired solution 
z(l) = x*. In other words, one considers a family of smoothly changing 
problems, the final problem being the original problem in question and the 
initial problem being one whose solution is easily determined. Use of the 
continuation, therefore, requires an ability to solve a sequence of problems 
when the solutions to nearby problems are known. 

A related, but conceptually distinct, problem arises hi many applica- 
tions in science and engineering where the equation under consideration 
always includes a number of physically relevant, intrinsic parameters. In 
other words, the equation has the generic form F(x,p) = 0, where x belongs 
to some state space X and p varies in a parameter space P. In this setting 
it is rarely of interest to determine only a few specific solutions. Instead, 
the requirement is to follow paths on the solution manifold {(z,p) e X X 
P} F(x, p) = 0} in the space XxP of all state and parameter variables and 
thereby to detect specific features of the manifold that signify a change of 
behavior of the system under study. 

For the first of these problems, namely the computation of a specific 
solution of some question, two distinct continuation techniques are avail- 
able. The first involves the case when the path of solutions to the family 
is smooth, which in turn allows its representation as a solution of a 
differential equation. The second approach is based on homological rather 
than homotopic concepts and makes use of a numerical form of a result in 
algebraic topology (Speraer's lemma). This approach has been refonnu- 

78 



lated and is now considered principally in the form of algorithms involving 
piecewise-linear solution paths. Much of the recent research in this area 
concerns these piecewise-linear continuation methods and their application 
to fixed-point problems in economics optimization and game theory. 

Continuation methods for following paths on the solution-manifold of 
parametrized equations developed mainly in structural engineering under 
the name of incremental methods. Applications to other areas, as, for 
example, to computational fluid dynamics and to phase transitions in 
statistical physics have only begun to appear relatively recently. For a 
numerical analysis of a given solution, manifold continuation methods have 
to be considered hi a broader sense as a collection of numerical procedures 
for completing a variety of tasks, including the following: 



1. Follow numerically any curve on the manifold specified by an a 
priori given combination of parameter values with one degree of freedom. 

2. On any such curve determine the exact location of target points 
where a given state variable has a specified value. 

3. K desired, at any such target point switch to the trace of a different 
solution path; 

4. On any such curve identify and compute the initial points where 
stability may be lost; 

5. From any one of the critical points follow a path in the critical 
boundary. In contrast to the case under 1, such paths are typically 
specified by combinations of the parameters with two degrees of freedom 
together with the implicit requirement that all points of the path are in 
the critical boundary. 

6. On any solution path determine the location of bifurcation points 
and the paths intersecting at that point. 



For specific applications additional tasks may arise. For instance, if 
the parametrized equation represents the equilibrium equation of a system 
of differential equations then we may wish to locate Hopf bifurcation points 
on a particular solution path where periodic solutions of the dynamical 
system branch off from the static equilibrium. 

In recent years much research has been devoted to these various tasks, 
but there remain many open questions especially hi connection with the 
more specialized tasks, such as the location of Hopf bifurcation points. 
Moreover, library-quality packages for the basic tasks 1-3 are still under 
development, and the software is by far not so advanced as, say, in the case 
of software for the solution of ordinary differential equations. For specific 

79 



applications, as, for example, to fluid-dynamic problems or to the case 
of structural problems involving plasticity, creep, and viscoelastic effects, 
the situation is still wide open; and only relatively ad hoc techniques are 
available. 



4.5 OPTIMIZATION METHODS 

Optimization problems occur in all areas of science, engineering, 
economics, and applied statistics. They may involve some least-squares 
approximation of observed data, fitting of parameters occurring in a math- 
ematical model on the basis of experimental observations, optimization of 
the design of an engineering structure, optimal control of an engineering 
system, or modeling of economic or business systems. These are only a 
few examples that do not even begin to cover the numerous types of op- 
timization problems that arise in practice. 

Broadly speaking, in all of these problems a real function, usually 
called the objective function, is to be minimized or maximized over some 
constraint set in a given finite- or infinite-dimensional space. The problems 
differ considerably depending on the mathematical characteristics of the 
objective function and the constraints, the dimension of the space, the 
amount of computable information that is available, and the requirements 
of specific applications. 

A good overview of the research problems concerning computa- 
tional methods for such optimization problems in finite-dimensional spaces 
has been given in the report Program Directions for Computational 
Mathematics, June 1979, prepared for the Applied Mathematical Sciences 
Research Program of the U.S. Department of Energy (edited by R. E. 
Huddleston, Sandia National Laboratories). This report identified the fol- 
lowing research areas: 



1. Analysis and comparison of techniques for dealing with nonlinear 
inequality constraints. 

2. Construction of algorithms for environments where computer 
storage is restricted. 

3. Production of high-quality software and other software-directed 
activities. 

4. Development of algorithms and software for problems in which 
some of the variables are discrete or integer. 

80 



5. Techniques to assist in finding global (or specified) minima. 

6. Large-scale linear programs with special structure, e.g., staircase 
and block-angular. 

7. Methods for nondifferentiable optimization. 

8. Techniques for selected nonlinear estimation problems, such as 
separable or constrained least squares, and data fitting in special norms. 

9. Investigation of the connection between fixed-point methods and 
alternative methods for solving nonlinear equations. 

10. Further study of selected aspects of linear constrained problems, 
such as special types of constraints, application of conjugate gradient 
techniques, the merits of various active set strategies, and the influence 
or scaling. 



The report does not address infinite-dimensional problems, for ex- 
ample, problems in control theory or the calculus of variations. Here the 
mathematical problems often become formidable, and special techniques 
are needed in most applications. For the computational solution, the prin- 
cipal approach applies finite-dimensional optimization methods to a dis- 
cretized form of the problem. This, in turn, raises the usual questions 
concerning the convergence and quality of the approximations. 



4.6 GRAPH-THEORETICAL METHODS 

Many physical systems, or their mathematical models, involve a number 
of discrete objects that are the producers or users of certain data or 
commodities and that in turn are interconnected by links that can transmit 
these items from one of the objects to another one. The underlying 
mathematical structure is then a graph consisting of a collection of vertices 
connected by arcs or branches. 

The modeling of some classes of physical systems in terms of graphs 
is rather natural, for example, in the case of communication or transpor- 
tation networks where the branches represent, say, power lines, roads, air- 
line routes, or water pipes, and, correspondingly, the vertices are power- 
generation stations, communities, airline terminals, or water reservoirs. In 
other cases the connection between a particular problem and its graph- 
theoretical interpretation is less obvious, for example, in the case of a 
discretization of a particular boundary-value problem on a given grid or 
the representation of a collection of data in a computer. 

81 



Corresponding to the wide range of applications, the forms of graph 
problems differ widely. For certain network models interest may center 
on the connectivity properties of the graph, that is, on the determination 
of whether a particular commodity can be transported between specific 
vertices. Then questions arise about the maximum possible flow that can 
be accommodated under particular constraints. These problems in turn are 
related to the so-called vulnerability and reliability problems for networks. 
On the other hand, if time enters into consideration then questions of 
waiting time and of best routing may arise. These are only a few of many 
types of such problems. 

From a computational viewpoint these various problems have stimu- 
lated the development of numerous combinatorial and graph-theoretical al- 
gorithms. But there remain many open research questions, especially when 
it comes to the production of general, robust software and the availability 
of algorithms for problems involving large graphs. 

Graph-theoretical formulations are also finding increasing application 
in connection with the numerical solution of problems that do not have an 
inherent graph-structure. One such class of problems concerns the com- 
putational handling of large sparse matrices. Many of the algorithms used 
here perform a sequence of steps that transform the matrix into succes- 
sively simpler matrices of, say, a more nearly upper triangular or diagonal 
type. These transformations achieve their aim by introducing zeros in 
place of originally nonzero matrix elements. But, as an unavoidable by- 
product they also introduce nonzero elements hi places where the original 
matrix elements were zero. Thus, in order to exploit the sparsity struc- 
ture of the matrix one must provide a data structure that either allocates 
from the outset space for all the fill-in during the computation or that 
allows for a dynamical allocation of space for the fill-in when it occurs. In 
both cases, graph-theoretical approaches form the basis for the design of 
desirable algorithms. 

An example of the use of a static data structure is the case when a row 
and column representation can bring the matrix into banded form with 
small bandwidth. This is frequently the case of the linear systems arising 
hi the finite-element discretization of elliptic boundary-value problems, 
and the corresponding bandwidth optimization routines are widely used 
in practice. On the other hand, if a dynamic storage structure is used, then 
special care has to be taken in the design of pivoting strategies for reducing 
fill-in while at the same time maintainmg numerical stability. Various 
sparse matrix packages have been developed for this purpose. They are 
typcially rather large and complex pieces of software. But the field is still 
under active development, and there remain many open research questions. 



82 



Iii particular, much still needs to be done on the interrelation between 
sparse matrix algorithms and special-purpose computer architectures. 

In recent years graph-theoretical approaches have also found increas- 
ing application in the design of algorithms for the numerical solution of 
problems of mathematical physics, in particular, of fluid dynamics. For ex- 
ample, it has been observed that certain natural finite-difference discretiza- 
tions of the equations of viscous, incompressible flow admit interpretations 
as systems defining flows on certain associated networks. Typically, the 
network nodes correspond to the idealized control volumes represented by 
the mesh points of the finite-difference equations, whereas the network 
arcs correspond to the paths on which one may identify the discrete finite- 
difference mass fluxes. Such observations can lead to remarkable savings 
in computing costs because they open the way to a priori transformations 
of the original (discretized) system of equations to completely equivalent 
systems in substantially fewer variables. For flow problems in two and 
three space dimensions the reduction factors are nominally 1/3 and 1/2, 
respectively. In graph-theoretical terms this approach corresponds to the 
transformation of the Maxwell-node equations to the Maxwell-mesh equa- 
tions, long used in electrical-circuit problems. 

The key requirement hi the application of network techniques to fluid- 
flow problems is that the discretized momentum and continuity equations 
be interpretable as "Ohm's laws" and "Kirchhoff-node laws" on an as- 
sociated network. This requirement permits great generality in the actual 
form of these laws and easily accommodates both implicit time-dependent 
as well as nonlinear steady-state difference forms of the Navier-Stokes 
equations. 

It is even possible to extend the network approach to compressible flow 
problems. Here the new idea appears to be the introduction of pseudo- 
flows into the node law and the identification of corresponding pseudo- 
characteristics to augment the Ohm's laws obtained from the momentum 
equations. In the nonstationary case, this produces numerical schemes 
whose stability is independent of local acoustic velocities. 

The entire area is still under active research development. For ex- 
ample, there are efforts to bring multispecies flow problems under the 
purview of network theory. This includes, as an important special case, 
two-phase flow problems that are a central concern in the design of reliable 
energy-generating systems and the safety analysis of nuclear power plants. 



83 



4.7 SPLITTING METHODS AND DEFECT CORRECTIONS 

Splitting is a means to separate a large intractable problem into a sequence 
of smaller, or at least more easily solved, problems. These methods are 
invoked to reduce significantly the computational effort (time and memory) 
needed to solve a problem compared with solving it directly. Often a 
form of splitting is required to make solution practical, and some cases 
even tractable. Various forms of splitting are common in engineering and 
scientific applications, even if they are not always recognized as such. 

Defect correction is a widely used, if often unlabeled technique of 
splitting. It presumes that one wants to solve a given hard problem, that 
one has in hand a guess at its solution, and that one also has a nearby 
problem that can be solved easily. It corrects the guess by solving the 
easier problem with special data computed from the guess. 

To illustrate the method, suppose that after discretization it is neces- 
sary to solve a finite-dimensional system of equations of the form 



- 

where A is a nonlinear operator, 6 is a known vector, and v is the solution. 
Often v is difficult to obtain directly, but the residual error 

r A(w) b 

for an approximate solution w is easy to evaluate. If there is a related 
system 



that approximates the original system and that is easier to solve, the defect 
correction algorithm may be appropriate. Given an estimate v n near a 
root v n+ i of the original system, we can expand the original equation in 
a Taylor series to get 

AK) - 6 + P(v n+1 ) - P(v n ] - ( J P - J A )(v n+l - v n ] + 0(e 2 ) 

where e = v n+ i v n . The defect correction iteration is any approximation 
to the above equation. The simplest such iteration is 



This iteration will converge if v n and J P (the Jacobian of P) are near 
enough to v n+i and J At respectively. 



84 



One of the most common splitting algorithms in engineering applica- 
tions is the alternating direction implicit (ADI) method. Here, the model 
problem is a partial differential equation system having two or more in- 
dependent spatial coordinates. Its direct solution requires too much com- 
puter storage and time to be tractable. Using an ADI splitting, the prob- 
lem is reduced to a manageable sequence of one-dimensional problems. 
In terms of the defect corrections, the system P is the sequence of one- 
dimensional problems that is much easier to solve, yet that closely ap- 
proximates the original system. 

Other forms of splitting also arise. For example, in certain systems of 
partial differential equations some of the equations are weakly coupled to 
the others. In these cases, solving the equations sequentially (rather than 
coupled) can result in significant savings. Similarly, in some combustion 
problems, considerable savings are realized through operator splitting al- 
gorithms in which the chemical rate terms are handled separately from the 
species transport terms. These methods are equivalent to matrix splittings 
of the Jacobian of the system. 

Analysis of splitting methods is important because the methods often 
do not converge. We must be concerned about accuracy and convergence 
of the factored system. Even though each iteration may be fast, we 
must have some idea about the overall convergence rates and about any 
degradation in accuracy resulting from the split. Analysis will also likely 
lead to the identification of matrix properties that suggest a beneficial 
splitting that would not be apparent from physical reasoning. Splitting 
can be considered as either a splitting of the original equations or as an 
approximate factorization of the iteration matrix. The former is the most 
intuitive, and the one in which physical insight is valuable. However, the 
latter is probably the one more amenable to analysis. 



4.8 MONTE CARLO TECHNIQUES 

Mathematical solution methods can be broken into two broad classes, 
Monte Carlo methods and deterministic methods, depending on whether 
chance events are included in the method. All scientists are familiar 
with deterministic methods but most have little familiarity with Monte 
Carlo methods. Deterministic methods do not depend on chance, and 
calculations performed using the same input data and the same computer 
code will provide exactly the same results. Monte Carlo calculations using 
the same input data and the same computer code will provide different 

85 



results, depending on what chance events occur. In this section we discuss 
what Monte Carlo methods are and how they might be improved. 

The Monte Carlo method estimates averages on a probability modd. 
Any quantity that can be expressed as an average value can be calculated 
by Monte Carlo techniques. Sometimes a probability model is obvious 
from the problem itself. For example, the probability that tossing three 
fair coins and obtaining a desired outcome [for example two heads (H) and 
one tail (T)] is easily estimated by Monte Carlo methods. The probability 
model consists of assigning a probability of ^ to an H and a probability 
of to a T on each toss and assigning a score of 1 to a desired outcome 
(HHT, HTH, THH) and a score of to any other outcome (HHH, TTT, 
THT, HTT, TTH). The computer has a random number generator that 
generates random numbers uniformly on the unit interval (0,1). A uniform 
distribution means that any random number is equally likely to have any 
value between and 1. Thus a com toss can be simulated on a computer 
by: 

H occurs if < 5 
T occurs if > f 

To toss three coins the computer uses three random numbers &, &, 
and 3 . A typical set of tosses might be i = 0.7 (T), & = 0.1 (H), and 
3 = 0.4 (H), scoring 1. The probability of obtaining two heads and one 
tail is (approximately) the average score (|) after many sets of computer 
tosses. 

Sometimes a probability model is not immediately apparent, but after 
a little thought the desired calculation can often be expressed as the 
estimation of an average value. For example, the integral 



can be thought of as the average value of f(x) on the interval (a, b). To 
see this, note that by definition the average value of f(x) with respect to 
a probability density p(x] is 



f b 
</>=/ f(x)p(x)dx. 

Ja 



Thus J is the average value of f(x) with respect to the probability density 
p(x} = l/(b a). Here p(x) is a uniform probability density on (a, b), and 



86 



x values can be samples uniformly on (a, 6) by setting 



whereupon 

i N 



where I is the estimate of I. . , , . , 

There are many applications for the Monte Carlo technique, but the 
application to neutron and photon transport has probably consumed more 
computer time than all other applications combined. Because ol tms, ana 
because neutron and photon transport problems have a natural pro baouiy 
model, the remaining discussion will pertain to neutron or photon ^trans- 
port problems. Furthermore, because neutron and photon transport are 
similar, the remaining discussion will refer to particle transport. 

Particle flow in nature is a random process. One cannot say exactly 
what will happen to an individual particle. One can only say what the 
probabilities are that various events occur. For example, a particle ttavei- 
ing through matter has a certain probability to collide per unit s , 
the actual distance between collisions is unknown, but the probab iihty ol 
traveling a distance I without collisions is known. Smulariy the ^nucuae 
a particle will collide with is not known, but the probability of colliding 

with the nuclide is known. ^,Wo-mQ ns 

The simplest Monte Carlo model for particle transport P^lems uj 
the natural probabilities that various events occur for the probabib 
model in essentially the same way as the coin toss example. That 
particles are followed from event to event in a ComputerLand tne 
is always selected (using the random number generator) trom a 
possible next events according to the natural event ****** 
type of model is called the analog Monte Carlo model because it is 
analogous to the naturaUy occurring ^transport 



particles contribute, and the statistical uncertainty m the answer 



models for particle transport that will estimate the same average value 
as the analog Monte Carlo model. These other techniques are known 
as nonanalog Monte Carlo models, and they are useful because although 
the average value remains unchanged, the variance (uncertainty) of the 
estimate can often be made much smaller than the variance for the analog 
estimate. Practically this means that problems that would be impossible 
to solve in days of computer time can be solved in minutes of computer 
time. 

A nonanalog Monte Carlo model attempts to follow "interesting" par- 
ticles more often than noninteresting particles. An "interesting" particle, 
by definition, is a particle that contributes a large amount to the quantity 
(or quantities) that needs to be estimated. There are many nonanalog 
techniques, and they all are meant to increase the odds that a particle 
scores (contributes). To ensure that the average score is the same in the 
nonanalog game as in the analog game, the score is modified to remove 
the effect of biasing (changing) the natural odds. Thus, if a particle is 
artificially made q times as likely to execute a given random walk, then 
the particle's score is weighted by (that is, multiplied by) 1/q. The average 
score is thus preserved because the average score is the sum, over all ran- 
dom walks, of the probability of a random walk and the score due to that 
random walk. The trick in obtaining low- variance solutions is to choose g's 
such that all random walks contribute the same score, in fact the average 
score. This then would be a zero-variance solution. 

It is always possible for any (linear) particle transport problem to 
select q's for each random walk such that every particle's score is the 
average score; that is, a zero-variance solution. The hooker is, of course, 
that the perfect q'& are not known, thus a zero-variance solution is not 
practical. However, people have often been successful enough in guessing 
good g's, that is biasing the odds, to improve the calculational efficiency 
by several orders of magnitude. This is typically done iteratively with a 
person making several short Monte Carlo computer runs (calculations) and 
using information gained on each run to better guess how to bias the next 
run. When the person's guesses no longer improve the calculation, a long 
run is done with the optimal biasing learned in the short runs. 

Can the computer learn to optimize the biasing? The computer is, 
after all, supplying the information that the person uses to learn. In 
the past decade a number of computer learning techniques have been 
tested. Thus far it has proven impossible for a computer to take an 
arbitrary transport problem and, without human aid, optimize the biasing. 
However, two things are worthy of note. First, computer learning aided 
by human judgment appears to be substantially better in many cases than 
human learning alone. This typically works by having the computer inform 

88 



the person how the computer would bias the subsequent run and having 
the person selectively alter the computer's suggestions according to the 
person's intuition. Second, the amount of human judgment required is 
decreasing. The day may come when the computer needs no human aid. 
Once human aid is no longer needed, the computer can learn to 
adjust its own biasing particle, as the calculation proceeds. An interesting 
implication of an adaptive Monte Carlo technique is that the common 
central limit theorem of statistics that would constrain the accuracy of 
the calculation to decrease as the square root of the number of particles 
followed no longer applies. The common central limit, theorem applies 
only when each particle's random walk is independent of all others and 
the sampling process is identical for all particles. Consequently, the rate 
of convergence may be more rapid. Although some limited statistical 
theories exist for dependent random variables, it has not been investigated 
whether these theories can profitably be applied to Monte Carlo particle 
transport problems. Thus, it is uncertain how fast an adaptive Monte. 
Carlo calculation is converging or indeed what the maximum convergence 
rate for a good learning process might be. Empirical evidence shows that 
the convergence can be substantially faster than the square root of the 
number of particles. In light of this, Monte Carlo methods can be expected 
to become more competitive with deterministic calculations. 



4.9 PROBLEM-DEPENDENT METHODS 

A variety of adaptive methods have a common goal: to overcome the 
computer size and grid resolution limitations, which are especially severe 
in singular problems, by use of computational elements that fit or model 
the singularity more directly. In this way, it may be possible to incorporate 
into the solution algorithm considerable analytic information beyond that 
provided by the equations themselves. This theme occurs in many aspects 
of numerical analysis. In the finite-element method, one may choose 
elements that include any singularities in the solution being computed (see 
Section 3.3). Grid adaptivity (see Section 4.2) is also problem dependent. 
Here we discuss examples of somewhat more special methods. Of necessity 
they apply only to classes of problems that possess related singularities. 



89 



4.9.1 The Riemann Problem and Nonlinear Wave Modes 

The Riemann problem for a nonlinear hyperbolic system of conservation 
laws is the Cauchy problem in one space dimension for data that are 
everywhere constant except for a single jump discontinuity at the origin. 
The solution of the Riemann problem provides a resolution of this discon- 
tinuity into the nonlinear modes, or waves, which propagate coherently 
in time. This idealized problem can be thought of as an approximate 
description of a higher-dimensional flow field especially in the neighbor- 
hood of a discontinuity hypersurface. This point of view has led to a num- 
ber of improved finite-difference schemes, which attempt to determine the 
various nonlinear wave modes at each point in space and time and to ad- 
just the differencing of the differential equations to take advantage of this 
knowledge. This adaptivity is to the structure of state space, in contrast 
to the coordinate space adaptivity discussed in Section 4.2. 



4.9.2 Front Tracking 

Front tracking is a combination of adaptive grid methods with the use of 
Riemann problems. The method is adapted to problems that have impor- 
tant singularity hypersurfaces (lines in two space dimensions, surfaces in 
three space dimensions), such as shock waves and contact discontinuities. 
In one version of this method, there is an overall time-dependent coordinate 
transformation to map the discontinuity into a fixed grid line in some set 
of computational coordinates. In another version, the discontinuity is rep- 
resented by a moving lower-dimensional grid that follows ("tracks") the 
dynamical motion of the discontinuity. The motion of the discontinuity 
and of the moving grid points that track it are governed by solutions of 
Riemann problems, or equivalently by the method of characteristics. 



4.9.3 Vortex Method 

This method introduces small lines or surfaces of vorticity into a fluid flow. 
The method is adapted to the study of shear-layer discontinuities, bound- 
ary layers, boundary-layer separation, and turbulence. The equations of 
motion of an ideal fluid yield simple equations for the motion of a collec- 
tion of vortices imbedded in the fluid. In fact, the vortices move passively 

90 



with the fluid, and their mutual interaction is described by a Hamiltonian 
system of ordinary differential equations, with Coulomb type interaction 
energy. La the case of the Navier-Stokes equations, the vortex motion also 
contains a diffusion term. 

These methods have been applied successfully to the problem of tur- 
bulent flame propagation (see Section 2.2). In this problem, the turbulent 
mixing is a primary factor in determining the flame speed. The turbulence 
comes from the boundary layer and in the boundary layer is calculated by 
a vortex method. Related methods have been developed under a variety of 
names such as boundary integral methods and Green's function methods. 



4.9.4 Scale Invariance and the Renormalization Group 

Scale transformations are the transformations x * sx, acting on space or 
on space and tune. A function u is homogeneous of degree a if 

u(sx, ay, sz) = s a u(x, y, z) 

and scale invariant if a = 0. Many problems have solutions that are scale 
invariant or approximately scale invariant over some parameter range. 
Such solutions are called similarity solutions. Using the scale invariance, 
one of the independent variables can be eliminated, making the solution 
more elementary to compute. 

However, scale invariance can also indicate the occurrence of com- 
plex phenomena. Specifically, any singularity that occurs in a scale- 
invariant problem must be repeated in all length scales (for which the 
scale invariance holds). Mathematically, Cantor sets, snowflake curves, 
and fractals are examples of such phenomena. In statistical physics, criti- 
cal phenomena in the equation of state is a scale-invariant problem. One 
general picture of turbulence holds that scale invariance (vortices on a 
large range of length scales) describe the inertia! range, or energy transport 
range, of turbulence. 

To implement these ideas in a computational algorithm, one integrates 
over a given set of length scales in a standard manner. The result of this 
computation is taken as data for a new calculation over a new set of length 
scales (with the original degrees of freedom eliminated from the prob- 
lem). This process is iterated and if convergent gives the scale exponent 
a, Sample numerical calculations of this type were discussed in Section 
2.1 in connection with turbulence. The method is well established for per- 
turbative calculations (with a small parameter) in critical phenomena in 
statistical physics. 

91 





25 




10 




30 




15 




35 




20 




37.5 



FIGURE 4.1 The evolution and merger of isolated vortex structures as 
predicted by contour dynamical techniques. [From E. A. Overman n and 
N. J. Zabusky, Phys, Fluids 25, 1297-1305 (1982).] 



92 



Since the renormalization group methods are novel, it is worth men- 
tioning that the mathematical foundations of this method have been estab- 
lished recently in several cases including examples of hierarchical models 
in statistical mechanics, interval maps, and renormalized quantum fields. 



4.9.5 Contour Dynamical Methods 

Contour dynamical methods are being applied to a variety of incompress- 
ible flows in two dimensions. These generalizations of the "waterbag" 
method provide simplified models for following the evolution of contours 
separating regions of piecewise-constant (pc) density that are the sources 
of the flow. The flow is the result of the self and mutual interaction of con- 
tours that evolve, mainly by area-preserving maps. These methods have 
been applied to the Euler equations, where pc finite-area-vortex regions 
and/or vortex sheets at density interfaces are sources of the flow; and 
the equations for a weakly ionized and strongly magnetized ionospheric 
plasma cloud, where pc ion-density regions are sources of the flow. For the 
former, a large class of steady-state translating and rotating solutions with 
pc vorticity ("V-states") have been found. Figure 4.1 shows the merger 
and breaking for a perturbed corotating V-state (two pc finite-area vortex 
regions with the same circulation), a familiar process in free-shear layer 
experiments. Notice how the two regions merge to form one region (by 
snipping out the common boundary at t = 10) and then stabilize by eject- 
ing vorticity in thin filaments. With these methods it is possible to obtain 
detailed information about the regions because the dimension is reduced 
by one. The curvature of the contour provides a predictive signature of 
the evolving small-scale structure, e.g., the roll up of vorticity filaments, 
etc. 



4.10 COMPUTER SOFTWARE AND HARDWARE 

As previously discussed, large-scale scientific calculations that tax the 
resources of the most powerful computers will continue to be essential to 
modern research and development efforts. To obtain long-term reliability 
and stability of future applications codes, implementing and testing of 
high-level numerical software should be coordinated. This will require a 



93 



strong research and development effort with cooperation supported among 
applications programmers, the theoretical numerical methods researchers, 
and the computer-science software developers. 

Repetition of expensive, error-prone, and time-consuming coding of 
commonly used methods should be avoided. Much of the current scientific 
software now being developed is redundant. If the common elements of the 
existing codes were available as modules, future applications programmers 
could use these routines and eliminate much of their efforts. New software 
is most readily accepted if it is compatible with existing techniques and 
simple enough so that potential users can observe tangibly better results 
in a trial run than those existing methods can produce. If such high- 
level routines were available, they could perform many of the common 
procedures found in applications codes, including grid generation, rezon- 
ing, numerical interpolation, differentiation, and integration; they could 
approximate differential boundary conditions and solve large, sparse non- 
linear systems of equations. 

An important goal is the machine independence of applications pro- 
grams. This is a difficult task because methods tailored specifically for a 
particular machine architecture will probably become more the norm than 
the exception. We can, by keeping machine-dependent codes in libraries of 
high-level software with standard user interfaces, strive to keep the user's 
scientific applications codes portable. The underlying library routines will 
be, necessarily, less portable because the architecture of the new machines 
will certainly be different from that of today's supercomputers. To uti- 
lize the inherent powers of the new machines we will have to re-examine 
traditional methods and identify the better ones for a particular machine 
architecture. 

The continuing revolution in microelectronics is having a profound 
impact on scientific computing. Indeed, it is likely to change our concept 
dramatically of what scientific computations can and cannot be effectively 
performed. Most certainly the impact of this revolution will be much 
greater than, say, the impact that floating-point hardware had. Moreover, 
while the costs of individual tasks will be greatly reduced, the domain of 
scientific computations will be greatly expanded and "frontier computa- 
tions" will continue to be expensive. 

These changes are being brought about by a number of factors: 
Individual components are becoming increasingly faster and smaller. Very- 
large-scale integration of such components is not only reducing the size of 
the packaged systems but also providing opportunities for customized, in- 
formation processors. Also, the decrease in the cost and size of computer 
memory implies that we can look for much larger memory systems. This 
will obviate many of the existing difficulties with secondary storage. 

94 



Despite the fact that components are becoming faster, the limits of 
raw machine speed are not visible, and further gains will eventually have 
to be made by using clever architectures and algorithms. Some kind of 
parallelism seems to be unavoidable. The programming issues involved in 
using parallel machines are still not satisfactorily resolved. The automatic 
detection of parallelism and the resulting scheduling of multiple resources 
are important open problems. 

Architectures, such as systolic arrays, based on specific subtasks can 
increase the performance of systems involving these subtasks by several 
orders of magnitude and clearly have a bright future. Algorithms contain- 
ing compute-intensive subtasks that can be vectorized in this fashion have 
a bright future. Because these architectures are hi general regular, the 
algorithms that can be vectorized for such machines tend to be regular, 
i.e., simple, nonadaptive, uniform-grid, low-order algorithms. It is clear 
that there are nicely behaved problems for which these regular algorithms 
on specialized machines will require significantly less time than algorithms 
requiring fewer operations on serial machines. It is also clear that no 
matter how fast the specialized machines are, there are problems that 
are sufficiently difficult that more sophisticated algorithms are needed for 
more general-purpose computers. 

In order to bring about these advances in architecture, it is necessary 
to involve practitioners of scientific computation in the design process. 
Luckily, modem design automation tools should make it possible for in- 
terdisciplinary design teams to successfully synthesize innovative special- 
purpose systems, and automated fabrication facilities should make it pos- 
sible for such systems to be built, debugged, and used. 

These advances in machine architecture should also have a dramatic 
effect on the design and analysis of algorithms for scientific computing. 
Traditionally, such analysis has been based on (asymptotic) estimates of 
the number of arithmetic operations. However, with these new architec- 
tures it is quite likely that the running time of an algorithm will be more 
dependent on the movement of data than on the number of arithmetic 
operations. Thus, we need to develop new analytic models of complexity of 
scientific algorithms so that such models give us useful information about 
the relative performance of algorithms. 



95 



Appendix A 
Persons Contributing to this Report 



Ivo Babuska, University of Maryland 

K. Binder, KFA Julich GmbH, Institut Festkorperforschung 

Norman Bleistein, University of Denver 

Thomas E. Booth, Los Alamos National Laboratory 

Billy L. Buzbee, Los Alamos National Laboratory 

William S. Childress, New York University 

Alexandre J. Chorin, University of California, Berkeley 

Lawrence D. Cloutman, Los Alamos National Laboratory 

James P. Corones, Ames Laboratory, Iowa State University 

Joel E. Dendy, Los Alamos National Laboratory 

Paul P. Eggermont, University of Delaware 

Alan J. Faller, University of Maryland 

Paul R. Garabedian, New York University 

John M. Greene, General Atomics Company 

Eugene Isaacson, New York University 

Jasper A. Jackson, Los Alamos National Laboratory 

Malvin H. Kalos, New York University 

Herbert B. Keller, California Institute of Technology 

Heins-Otto Kreiss, California Institute of Technology 

Edward W. Larsen, Los Alamos National Laboratory 

Peter D. Lax, New York University 

Joel Lebowitz, Rutgers University 

Dennis R. Liles, Los Alamos National Laboratory 

C. C. Lin, Massachusetts Institute of Technology 

Oscar Manley, U. S. Department of Energy 

Steven B. Margolis, Sandia National Laboratories 

Jerrold E. Marsden, University of California, Berkeley 

Oliver A. McBryan, New York University 

Keith Miller, University of California, Berkeley 

Cathleen S, Morawetz, New York University 

M. Zahair Nashed, University of Delaware 

J. Tinsley Oden, University of Texas 

Edward A. Overman H, University of Pittsburgh 

Edward Ott, University of Maryland 

96 



George C. Papanicolaou, New York University 

Charles S. Peskin, New York University 

Linda R. Petzold, Sandia National Laboratories 

Thomas F. Porsching, University of Pittsburgh 

Garry H. Rodrigue, Lawrence Livermore National Laboratory 

Jacob T. Schwartz, New York University 

Philip E. Seiden, IBM Corporation 

David H. Sharp, Lawrence Livermore National Laboratory 

Lawrence A. Shepp, Bell Laboratories 

Mitchell D. Smooke, Sandia National Laboratories 

Blair K. Swartz, Los Alamos National Laboratory 

Yvain M. Treve, La Jolla Institute 

Michael Steenstrup Vogelius, University of Maryland 

Burton Wendrqff, Los Alamos National Laboratory 

Norman J. Zabusky, University of Pittsburgh 



97 




511.3 N27c c.l 

National Research Council 

(U.S.). 

Computational modeling and 

mathematics applied to the 



University Libraries 
Carnegie- Mellon University 

Pittsburgh, Pennsylvania 15213 



DEMCO